US20150236957A1 - System and method for efficient transport of large data files - Google Patents

System and method for efficient transport of large data files Download PDF

Info

Publication number
US20150236957A1
US20150236957A1 US14/628,102 US201514628102A US2015236957A1 US 20150236957 A1 US20150236957 A1 US 20150236957A1 US 201514628102 A US201514628102 A US 201514628102A US 2015236957 A1 US2015236957 A1 US 2015236957A1
Authority
US
United States
Prior art keywords
chain
data
frames
network
control packet
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US14/628,102
Other versions
US9800509B2 (en
Inventor
Ilijc Albanese
Thomas E. Darcie
Sudhakar Ganti
Stephen W. NEVILLE
Yagiz O. Yazir
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
UVic Industry Partnerships Inc
Original Assignee
UVic Industry Partnerships Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by UVic Industry Partnerships Inc filed Critical UVic Industry Partnerships Inc
Priority to US14/628,102 priority Critical patent/US9800509B2/en
Publication of US20150236957A1 publication Critical patent/US20150236957A1/en
Assigned to UVIC INDUSTRY PARTNERSHIPS INC. reassignment UVIC INDUSTRY PARTNERSHIPS INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: YAZIR, YAGIZ ONAT, ALBANESE, ILIJC, DARCIE, THOMAS E., GANTI, SUDHAKAR, NEVILLE, Stephen
Application granted granted Critical
Publication of US9800509B2 publication Critical patent/US9800509B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/12Avoiding congestion; Recovering from congestion
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/64Hybrid switching systems
    • H04L12/6418Hybrid transport
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/18End to end
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/19Flow control; Congestion control at layers above the network layer
    • H04L47/193Flow control; Congestion control at layers above the network layer at the transport layer, e.g. TCP related
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/19Flow control; Congestion control at layers above the network layer
    • H04L47/196Integration of transport layer protocols, e.g. TCP and UDP
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/22Traffic shaping
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/28Flow control; Congestion control in relation to timing considerations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/30Flow control; Congestion control in combination with information about buffer occupancy at either end or at transit nodes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L47/00Traffic control in data switching networks
    • H04L47/10Flow control; Congestion control
    • H04L47/36Flow control; Congestion control by determining packet size, e.g. maximum transfer unit [MTU]

Definitions

  • the present disclosure relates to networking and the future evolution of the Internet, where increased efficiency and decreased power consumption can be achieved through an overlay network specifically designed to handle very large data packets that are representative of present-day and future large file transactions.
  • Telecommunication networks are experiencing continued rapid traffic increase mostly driven by the proliferation of bandwidth-intensive applications.
  • the revenues for network operators are not growing at the same pace.
  • the industry must reduce the cost per bit transported and increase capacity. While increasing the deployed capacity is costly, improving traffic grooming and shaping becomes important to better utilize the capacity provided.
  • Traffic shaping policies are often used by ISPs to limit bandwidth costs especially when it comes to bulk data transfers. In spite of their widespread use, such policies often lead to significant end-to-end performance losses.
  • TCP and UDP the increasing proportion of large file transactions (e.g. VoD, IPTV, CDN content update and distribution and so on), places an increasing burden on the network hardware and capacity of current IP networks.
  • Optical Transport Network In parallel, recent standardization of transport architectures such as ITU-T G.709/G.872 Optical Transport Network (OTN) has provided a key step towards a more flexible and efficient transport infrastructure layer, enabling protocol agnostic transport services, allowing a more efficient use of deployed capacity and simplifying management operations.
  • This widely deployed standard provides well-defined mappings of most existing and future higher layer formats into a common underlying transport layer.
  • a method for transporting a data file comprising dividing the data file into a plurality of data frames; concatenating the plurality of data frames into at least one chain with a predefined transparency degree; associating the at least one chain with a control packet comprising information regarding at least a chain length of the at least one chain and the predefined transparency degree; scheduling transmission of the at least one chain along a predetermined path of a network based on the control packet; and transmitting each data frame along the predetermined path at a time derived from an estimated time-to-transmit parameter.
  • the method can further comprise concatenating the plurality of data frames into the at least one chain with one or more void frames.
  • Dividing the data file into a plurality of data frames can comprise determining data frame sizes based on a transport technology.
  • the data frame size can be selected such that mapping onto a transport layer is as seamless as possible.
  • the method can further comprise using a buffer having a buffer size proportional to a data frame size and a chain confirmation, such as a transparency degree and frame size of the at least one chain, to delay the at least one chain so as to interleave the at least one chain with a previously received chain or an incoming chain.
  • the maximum buffer size can be selected based a maximum data frame size and maximum transparency degree allowed for a chain.
  • the data file has a size of at least 100 MB.
  • the predefined transparency degree of the at least one chain is associated with a number of periods of the at least one chain having at least one timeslot occupied by at least one data frame and in some embodiments, the predefined transparency degree is associated with a number of void frames in the at least one chain.
  • the estimated time-to-transmit parameter can be included in the control packet and can be communicated to at least one node of the predetermined path and is associated with a delay between control packet arrival and chain reception commencement.
  • a system for transporting a data file comprising at least one source server comprising a software module configured to: divide the data file into a plurality of data frames; concatenate the plurality of data frames into at least one chain with a predefined transparency degree; and associate the at least one chain with a control packet comprising information regarding at least a chain length of the at least one chain and the predefined transparency degree.
  • the software module can be further configured to schedule transmission of the at least one chain along a predetermined path of a network based on the control packet and transmit each data frame along the predetermined path at a time derived from an estimated time-to-transmit parameter.
  • the at least one chain further comprises one or more void frames.
  • the plurality of data frames can comprise data frames of different sizes.
  • the predefined transparency degree of the at least one chain can be associated with a number of periods of the at least one chain comprising at least one timeslot occupied by at least one data frame. In some embodiments, the predefined transparency degree is associated with a number of void frames in the at least one chain.
  • the estimated time-to-transmit parameter can be included in the control packet and can be communicated to at least one node of the predetermined path and is associated with a delay between control packet arrival and chain reception commencement.
  • the system includes a network interface that communicates the at least one chain and wherein the network interface is an optical transport network or an Ethernet-based network.
  • a network node comprising a network interface configured to receive at least one chain that includes a plurality of data frames and void frames, or a portion of the at least one chain, wherein the network node transmits an updated estimated time-to-transmit parameter to a destination network.
  • the network interface can be configured to receive a control packet associated with the at least one chain and can be configured to extract an estimated time-to-transmit parameter from the control packet.
  • the updated estimated time-to-transmit parameter can be determined by adding an original estimated time-to-transmit parameter of the network node to an original buffering time of the network node and subtracting an estimated control packet processing time for the network node.
  • FIG. 2 illustrates a data frame assembled by multiple basic payload frames (BPFs).
  • BPFs basic payload frames
  • FIG. 3 illustrates integration of the proposed protocol in the current layered architecture.
  • FIG. 4 illustrates a simulation topology embodiment
  • FIG. 5 is a graph of normalized goodput for BFP over Ethernet.
  • FIG. 6 is a graph of normalized goodput for TCP.
  • FIG. 7 is a graph of normalized goodput for transactions ⁇ 100 MB.
  • FIG. 8 is a graph of delay per transaction for transactions ⁇ 100 MB.
  • FIG. 9 is a graph of average buffer size for transactions ⁇ 100 MB.
  • FIG. 10 is a schematic diagram of a representative optical transport network (OTN) switch modified to support a big file protocol (BFP).
  • OTN optical transport network
  • BFP big file protocol
  • FIG. 11 is a schematic diagram of another representative optical transport network (OTN) switch modified to support a big file protocol (BFP).
  • OTN optical transport network
  • BFP big file protocol
  • values, procedures, or apparatus' are referred to as “lowest”, “best”, “minimum,” or the like. It will be appreciated that such descriptions are intended to indicate that a selection among many used functional alternatives can be made, and such selections need not be better, smaller, or otherwise preferable to other selections.
  • Bandwidth demand is increasingly dominated by large transactions.
  • the present disclosure concerns embodiments of a method and system useful for a networking approach wherein large file transactions are handled separately from the large number of smaller transactions. Traffic shaping can then be used to improve link utilization and reduce the computational load placed on the network. Traffic shaping for the corresponding long hold times can benefit from a far less dynamic traffic mix.
  • large transactions are shaped into periodic concatenations of data frames which are scheduled for transmission over an end-to-end path using a lightweight scheduling protocol.
  • This Big File Protocol (BFP) shapes each transaction prior to transmission and executes the scheduling procedure on a per-flow basis.
  • BFP creates a tighter linkage (than TCP/UDP) between the application layer and resource availability at the physical layer (either Ethernet or OTN), made possible by the long push times of long file transfers.
  • This approach does not require network-wide coordination.
  • the disclosed methods and systems can be implemented over any packet switched network, and are entirely compatible with the current OTN infrastructure as well as with Ethernet-based networks, which are commonly used in data-center environments.
  • the disclosed methods and systems are able to overcome the limitations of traffic shaping on end-to-end performance of the network while allowing a more efficient utilization of network resources, and a significant increase in goodput (>40%), potentially reducing the cost per bit.
  • the big file protocol disclosed herein is a cross-layer transport protocol designed to handle large transactions (e.g. 100 MB) in a more efficient manner by shaping traffic and leveraging currently available transport layer functionalities to enable a better use of the deployed capacity and reduce the computational load placed on the network by large transactions.
  • the nodes implementing BFP shape traffic from each transaction into periodic, semitransparent concatenations of fixed-size data frames, referred to as chains henceforth.
  • Periodicity of each chain is obtained by modeling access to the transport resource as a periodic succession of fixed length timeslots in which a transaction is allowed access to an output channel at full capacity. Periodicity simplifies scheduling and stabilizes traffic variations resulting in stable network performance.
  • the semi-transparency of each chain, resulting from periodic access to the channel allows interleaving several flows using buffers to align the flows in time.
  • the transparency degree (TD) of a chain is defined as the period (expressed in number of timeslots) of the occupied slots in a periodic frame structure that represents capacity on a transport resource.
  • BFP differs from TDM systems in that timeslots are only used to model transport resources and regulate access on a per flow basis, and need not to be physically implemented in the underlying hardware.
  • Another difference with TDM systems is that the timeslots in TDM are somehow rigid and each TDM channel occupies a fixed amount of the available bandwidth. In BFP the amount of bandwidth occupied is decided on a per-transaction basis, resulting in a more flexible bandwidth usage.
  • the size of the BFP timeslots can be adjusted by varying the size of the data frames.
  • the fixed-size data frames are comprised of an integer number of Basic Payload Frames (BPF) assembled back to back as shown in FIG. 2 .
  • BPF Basic Payload Frames
  • BVF Basic Void Frames
  • Scheduling is the means by which contending requests are directed to transmit so as to minimize buffering required to efficiently pack the transport resource.
  • Admission control is the process used to grant or deny permission to contend.
  • a network of fixed capacity links is presumed to have been provisioned interconnecting widely separated data centers and this capacity is accessed only by BFP.
  • This deployed capacity may be in the form of dedicated transport capacity, whether this is Ethernet, Ethernet over OTN, or a mapping of BFP directly onto the OTN layer.
  • the data frames fit naturally within the underlying transport capacity.
  • BFP exploits a tight linkage between application and physical-layer resources, it can be described with reference to FIG. 3 .
  • Application programmers must see this as yet another variant of well-known socket-related actions, like TCP and UDP, as discussed later.
  • TCP and UDP pass through IP before transport
  • BFP passes through a parallel path to transport. This path manages admission control, scheduling, and mapping into physical-layer resources, functions that may exploit information passed between nodes by TCP/IP.
  • CP control packet
  • ACK confirmation
  • an Expected Time of Arrival (ETAsrc) parameter is computed according to the following Equation (1) and included in the CP.
  • N number of nodes in the path ⁇ i propagation time to reach node i p j estimated CP processing time for node i ⁇ ack time for the ACK to reach the source node
  • the ETA parameter communicates to each node in the path the delay between the reception of the CP and the arrival of the first bit of its relative chain.
  • Each node uses ETA and chain configuration information to compute the buffering time (BT) to interleave an incoming chain with previously scheduled chains on the output channel, according to Equation 2.
  • BT buffering time
  • CP is updated with a new ETA information, computed according to Equation 3. This updated ETA value will be used by the next node in the path to schedule the chain.
  • an acknowledgement is generated and sent over the reverse path to the source node.
  • the source node Upon receiving the ACK, the source node is assured that resources are available to support the transaction and starts transmitting the data frames separated by a number of void frames (BVF) consistent with the selected TD.
  • BVF void frames
  • each node in the path can reserve resources only for the time necessary for the chain to traverse it, optimizing resource utilization. Buffering can be used to align frames in time and to overcome any timing uncertainties. Therefore, timing precision required by the scheduling procedure should be manageable using standard ranging procedures.
  • Standard Ethernet frames bare a payload varying from a minimum of 42 bytes to 1500 bytes. This limitation is removed when jumbo frames are used, allowing up to 9000 bytes of payload.
  • Each Ethernet frame has a preamble of 7 bytes, 1 byte of Start Frame Delimiter (SFD), 6 bytes for source MAC address and 6 for destination MAC address, a 4 byte Frame Check Sequence is also appended to the Ethernet frame. Furthermore, after each frame is sent, transmitters are should transmit a minimum of 12 bytes of Inter Frame Gap (IFG).
  • IFG Inter Frame Gap
  • BPF and BVF can be mapped directly onto Ethernet frames by setting the size of BPF and BVF equal to the Ethernet frame payload size (e.g.: 9000 B), allowing the proposed protocol to be mapped onto an Ethernet transport layer.
  • frame delineation is performed by the Ethernet frames and preamble or Inter Frame Gap (IFG) bits can be used to code frame-related information (e.g. frame number, frame type—BPF, BVF or CP— and data frame size).
  • IFG Inter Frame Gap
  • ITU-T G.709 provides bit and timing transparent transport services. Standard containers for any client signal available today are defined together with their relative mapping procedures. Furthermore, flexible containers (i.e. ODUflex) are defined in the G.709 transport hierarchy to support packet-based clients with a wide range of bitrates. Each OTN frame bares a payload of 15232 bytes over which one or more client signals are mapped. For cell or packet based clients OTN uses GFP to encapsulate data packets and generate a continuous stream of GFP frames which is then mapped in an octet-aligned manner directly onto the OPU payload area. In this case, rate adaptation is done using GFP idle frames, which are transmitted anytime there is no data to send.
  • the general idea is to reserve an ODU channel over each link of a network to build an overlay network which is then used exclusively to handle BFP transactions.
  • the portion of dedicated bandwidth is flexible.
  • setting up ODUflex channels for each link of the overlay network seems the best option as these channels can be resized without tearing down and re-establishing the connection via the Hitless Adjustment of ODUflex protocol.
  • the configuration of the overlay network should be relatively stable (time-wise) and only occasional changes will be needed. Such changes can be planned offline over coarse timescales.
  • each chain is built by mapping BPF and BVF onto GFP frames, and GFP Extension Header is used to code frame-related information.
  • the resulting stream of GFP frames is then mapped onto the payload area of the ODUflex (i.e. ODUflex(GFP)) signal in an octet-aligned manner. Frame delineation is left to GFP and any eventual rate adaptation is performed using GFP idle frames.
  • ODUflex i.e. ODUflex(GFP)
  • routing and load balancing algorithms can be specifically designed for the disclosed methods, taking into account the periodic configuration of the proposed transport structure, it is also possible to simply re-use protocols that are already in place. In the following a method to reuse deployed routing protocols is discussed.
  • each table entry is relative to a specific chain and is stored in the table until the chain traverses the node.
  • the size of this table depends on both on the number of transactions each node is able to handle and on the average duration of each reservation procedure. Although the size of the table may become cumbersome the assumption here is that large transactions are relatively infrequent and the reservation procedure is fast enough to keep the size manageable.
  • the protocol can be installed in the network stack in a way that provides a certain level of familiarity for the application developers. That is, an application developer, given the drive to use the protocol and the familiarity with the usage of network sockets, should be able to undertake the well-known and common socket related actions—such as creating, connecting, listening, sending, receiving, etc. —in a way that is not much different from the actions that are involved in programming with TCP or UDP sockets.
  • the proposed protocol can be registered with the socket layer, can be easily accessed by just using an indicator that will point to the newly registered protocol upon socket creation, and the functionality is provided with identical function headers. In the same sense, the structural similarities will also have a positive impact on incorporating the proposed protocol with software packages that abstract the socket layer.
  • the proposed BFP can be registered with the socket layer as a new transport layer protocol making it appear to the application developers and the abstraction packages as just another protocol like TCP and UDP, hence, hiding the cross-layer behavior of the proposed protocol at the lower layers (See FIG. 3
  • implementation of the proposed protocol in the form of a loadable kernel module (LKM) will ensure that the kernel patches, hence possible regressions, are avoided, and the new protocol can be unloaded upon request. This approach also ensures that the deployment is relatively easy from the network administrators' point of view.
  • LLM loadable kernel module
  • FIG. 10 An exemplary BFP ( FIG. 10 ) shows how BFP functionalities can be integrated into an OTN-enabled NIC card.
  • Incoming optical signals (OTUk) are converted into electronic format and processed by the OTN processor (block 1000 ) before being passed to the Fabric Interface Chip (FIC, block 1002 ) as packets.
  • FIC Fabric Interface Chip
  • BFP traffic is separated from non-BFP traffic.
  • BFP Control Packets (CP) are processed and information about configuration and timing of their relative chains is used by the FIC to schedule buffer space and switching resources for the data chains.
  • CP source and destination addresses are passed to the routing layer through the Routing Layer Interface (block 1004 ) which communicates with whichever routing protocol is available and presents the FIC with an input-port-to-output-port mapping relative to the addresses provided. This information is used by the FIC to populate a table (referred to as the port-map table), which contains the binding of input-to-output ports for each chain.
  • the Routing Layer Interface block 1004
  • This information is used by the FIC to populate a table (referred to as the port-map table), which contains the binding of input-to-output ports for each chain.
  • Data frames from each chain are routed through a buffering stage (block 1006 ) to be delayed by an amount of time consistent with the buffering time (BT) computed for each chain.
  • BT buffering time
  • chains can be sent to their respective output ports through the switch fabric. Chains going through the same ports will naturally interleave with one another.
  • Block 1002 can also be implemented with two separate blocks as is illustrated in FIG. 11 , namely: FIC ( 1102 , FIG. 11 ) and Traffic Manager and Packet Processor ( 1100 , FIG. 11 ), whose functionalities are as follows:
  • Traffic manager (TM)+packet processor (PP) illustrated as block 1100 in FIG. 11 : the traffic manager separates BFP traffic (control and data) from standard traffic (i.e., non-BFP traffic).
  • CPs are processed in this block and BT is computed for each incoming chain in advance.
  • this block requests routing information to the routing layer through the routing layer interface ( 1004 , FIG. 11 ).
  • the TM schedules switching resources and reserves buffer space (in advance) for incoming chains.
  • the TM also populates/updates the port-map table with the newly obtained routing information.
  • BFP data frames i.e., chains
  • these are routed to the buffer stage ( 1006 , FIG. 11 ), where each data frame is delayed by an amount of time consistent with its relative BT before being routed toward the FIC ( 1102 , FIG. 11 ).
  • Fabric Interface Chip (FIC, 1002 of FIG. 10 / 1102 FIG. 11 ) usually involve the following: Fabric interface chip (FIC, 1002 of FIG. 10 / 1102 FIG. 11 )—this block performs standard FIC functions (i.e., traffic adaptation and congestion control) for both BFP and non-BFP traffic.
  • FIC functions i.e., traffic adaptation and congestion control
  • the ability to efficiently handle bulk data transfers of the disclosed method was evaluated in comparison to that of TCP Westwood, a high-speed version of TCP currently implemented in the Linux kernel. All simulations were run using Omnet++ discrete event simulation tool. The two approaches are compared in terms of goodput, end-to end delay and average buffer size.
  • a dumbbell topology ( FIG. 4 ) was selected for this example. The propagation delay for each link is set to 1 mS giving a RTT of roughly 6 mS. Bitrate was set to 10 Gb/s for all links.
  • the size of each data frame is set to be one BPF, and the same size is selected for the voids (i.e. one void is comprised of a single BVF) for all BFP cases. Transactions varying in size from 500 kB to 1 GB were considered and each transmission repeated until statistical stability is reached.
  • each OTN frame carries 15222B of payload, corresponding to the payload area of the OTN frame minus the GFP overhead including core, payload header, and two bytes of the extended header which are used as a binary counter to number each frame in a chain, allowing a maximum size for the bulk data carried in a chain of roughly 1 GB.
  • An Ethernet jumbo frame with 9000B of payload was selected as the basic payload frame for the BFP over Ethernet case.
  • a TD of 8 was used for all BFP cases and a maximum buffer size of 12 frames (TD+4) was selected for both BFP cases. This is not a strict requirement and can be relaxed by trading buffer space for higher utilization and lower CP blocking probability.
  • a MSS of 8960B was selected and the layer 2 MTU was set to 9000B (Ethernet jumbo frame).
  • FIGS. 5 and 6 show the average normalized goodput for BFP over Ethernet and TCP respectively. Performance of BFP over OTN were also studied. Results show nearly identical performance of BFP over OTN and over Ethernet. Due to space constraints only the latter is shown here.
  • each source tends to fill the entire bandwidth of their link to the bottleneck router which gives higher goodput with respect to the BFP case for long lived TCP flows ( ⁇ 300 MB) as long as the number of sources remains limited ( ⁇ 3).
  • TCP either does not ramp up fast enough to fill the available bandwidth or, as the number of sources increases, TCP is prone to packet drops which trigger the TCP backoff algorithm, reducing the offered load from the TCP sources and resulting in poor link utilization.
  • the average time needed to successfully complete one transaction was selected, including the time needed to setup the connection and any data retransmission.
  • a new connection is setup for each new transaction.
  • a comparison of the end-to-end delay performance for transactions ⁇ 100 MB is shown in FIG. 8 .
  • TCP tends to fill the available bandwidth and achieves smaller delays.
  • TCP goodput collapses see FIG. 6
  • end-to-end delay rapidly increases to values up to over 30 times larger than BFP. Similar performance can be seen for all other transaction sizes (e.g. 300 MB and 100 MB), although in these cases BFP achieves better delay performances over the entire range considered.
  • TCP delay performance
  • ACK acknowledgements
  • RTT Round Trip Time
  • Delay will further degrade in the TCP case as the network diameter increases. This would also happen in the BFP case but to a much smaller degree due to the much smaller amount of control information exchanged between source and destination.
  • the particular buffering technique used by BFP also improves delay performance with respect to TCP by reducing the queuing delays when the network is congested (See FIG. 9 ).
  • the variation of the end-to-end delay over the entire range considered is much smaller for BFP, resulting in more stable delay performance of BFP over TCP.
  • FIG. 9 shows a comparison of the average buffer space utilized by BFP and TCP for transactions ⁇ 100 MB.
  • Average buffer size for BFP is proportional to the BFP frame size and the selected TD, resulting in stable and predictable buffer occupancy in any load condition. The random nature of TCP packet arrival will tend to fill the buffer quickly as congestion approaches.
  • FIG. 9 shows that for a transaction size of 1 GB, buffer size grows rapidly to values several times larger than in the BFP case when more than 2 TCP sources compete for the same bottleneck link. Similar performance is observed for other transaction sizes.
  • the maximum buffer size for BFP is a function of the chains configuration rather than of the network load, and can therefore be tightly controlled.
  • the present disclosure concerns a BFP, a network protocol designed to efficiently handle large transactions over the existing network infrastructure.
  • BFP large file transactions are handled at lower layers (e.g. L1 and L2). Higher layers are accessed only when needed with routing functionalities that are only used during connection setup and not for every single data frame as in TCP. Scheduling allows BFP to accommodate network loads over 40% higher with respect to TCP without incurring goodput collapse and achieving goodput values close to 100%.
  • Stable end-to-end delay performance is also achieved, with values over 30 times smaller with respect to TCP.
  • Buffer occupancy of BFP in congested scenarios is predictable and much smaller relative to TCP.
  • the results presented show the potential of BFP to achieve better use of the deployed capacity, potentially resulting in significant cost savings for the network operators in both transport and data center environments.

Abstract

Telecommunication networks are under stress due to rapid traffic increase cause mostly by large file transfers. Disclosed herein is a cross-layer transport protocol specifically designed to efficiently handle large transactions. Traffic generated from large transactions is shaped into a periodic succession of fixed-size data frames. Each transaction can then be scheduled for transmission using a two-way reservation protocol. Exemplary results show that the proposed approach is capable of significantly improving goodput and end-to-end delay relative to TCP, improving efficiency of bandwidth utilization by over 40%.

Description

    CROSS REFERENCE TO RELATED APPLICATION
  • This application claims the benefit of U.S. Provisional Patent Application No. 61/942,552, filed Feb. 20, 2014, which is incorporated herein by reference in its entirety.
  • FIELD
  • The present disclosure relates to networking and the future evolution of the Internet, where increased efficiency and decreased power consumption can be achieved through an overlay network specifically designed to handle very large data packets that are representative of present-day and future large file transactions.
  • BACKGROUND
  • Telecommunication networks are experiencing continued rapid traffic increase mostly driven by the proliferation of bandwidth-intensive applications. Unfortunately, the revenues for network operators are not growing at the same pace. To continue supporting the growth of Internet-based applications in an economically viable manner, the industry must reduce the cost per bit transported and increase capacity. While increasing the deployed capacity is costly, improving traffic grooming and shaping becomes important to better utilize the capacity provided. Traffic shaping policies are often used by ISPs to limit bandwidth costs especially when it comes to bulk data transfers. In spite of their widespread use, such policies often lead to significant end-to-end performance losses. Furthermore, constrained by traditional protocols such as TCP and UDP, the increasing proportion of large file transactions (e.g. VoD, IPTV, CDN content update and distribution and so on), places an increasing burden on the network hardware and capacity of current IP networks.
  • In the increasingly important Data Center (DC) environment, efficient intra-DC and inter-DC networks play a crucial role in minimizing congestion and conserving computational resources. These networks must provide predictable performance for diverse and unpredictable load conditions. Managing this unpredictability generally translates into inefficiency in the use of transport resources such that associated transport costs become appreciable, particularly for wide-area inter-DC networks. In this context, a network that is able to provide predictable network performance for random traffic patterns would be a desirable feature.
  • In parallel, recent standardization of transport architectures such as ITU-T G.709/G.872 Optical Transport Network (OTN) has provided a key step towards a more flexible and efficient transport infrastructure layer, enabling protocol agnostic transport services, allowing a more efficient use of deployed capacity and simplifying management operations. This widely deployed standard provides well-defined mappings of most existing and future higher layer formats into a common underlying transport layer.
  • SUMMARY
  • Disclosed herein are embodiments of a method for transporting a data file, comprising dividing the data file into a plurality of data frames; concatenating the plurality of data frames into at least one chain with a predefined transparency degree; associating the at least one chain with a control packet comprising information regarding at least a chain length of the at least one chain and the predefined transparency degree; scheduling transmission of the at least one chain along a predetermined path of a network based on the control packet; and transmitting each data frame along the predetermined path at a time derived from an estimated time-to-transmit parameter. In some embodiments, the method can further comprise concatenating the plurality of data frames into the at least one chain with one or more void frames.
  • Dividing the data file into a plurality of data frames can comprise determining data frame sizes based on a transport technology. For example, the data frame size can be selected such that mapping onto a transport layer is as seamless as possible. In some embodiments, the method can further comprise using a buffer having a buffer size proportional to a data frame size and a chain confirmation, such as a transparency degree and frame size of the at least one chain, to delay the at least one chain so as to interleave the at least one chain with a previously received chain or an incoming chain.
  • The maximum buffer size can be selected based a maximum data frame size and maximum transparency degree allowed for a chain. In some embodiments, the data file has a size of at least 100 MB.
  • The predefined transparency degree of the at least one chain is associated with a number of periods of the at least one chain having at least one timeslot occupied by at least one data frame and in some embodiments, the predefined transparency degree is associated with a number of void frames in the at least one chain.
  • The estimated time-to-transmit parameter can be included in the control packet and can be communicated to at least one node of the predetermined path and is associated with a delay between control packet arrival and chain reception commencement.
  • Also disclosed herein are embodiments of a system for transporting a data file, comprising at least one source server comprising a software module configured to: divide the data file into a plurality of data frames; concatenate the plurality of data frames into at least one chain with a predefined transparency degree; and associate the at least one chain with a control packet comprising information regarding at least a chain length of the at least one chain and the predefined transparency degree. In some embodiments, the software module can be further configured to schedule transmission of the at least one chain along a predetermined path of a network based on the control packet and transmit each data frame along the predetermined path at a time derived from an estimated time-to-transmit parameter.
  • In some embodiments of the systems, the at least one chain further comprises one or more void frames. The plurality of data frames can comprise data frames of different sizes. Also, the predefined transparency degree of the at least one chain can be associated with a number of periods of the at least one chain comprising at least one timeslot occupied by at least one data frame. In some embodiments, the predefined transparency degree is associated with a number of void frames in the at least one chain. The estimated time-to-transmit parameter can be included in the control packet and can be communicated to at least one node of the predetermined path and is associated with a delay between control packet arrival and chain reception commencement. In some embodiments, the system includes a network interface that communicates the at least one chain and wherein the network interface is an optical transport network or an Ethernet-based network.
  • Also disclosed herein are embodiments of a network node, comprising a network interface configured to receive at least one chain that includes a plurality of data frames and void frames, or a portion of the at least one chain, wherein the network node transmits an updated estimated time-to-transmit parameter to a destination network. The network interface can be configured to receive a control packet associated with the at least one chain and can be configured to extract an estimated time-to-transmit parameter from the control packet. The updated estimated time-to-transmit parameter can be determined by adding an original estimated time-to-transmit parameter of the network node to an original buffering time of the network node and subtracting an estimated control packet processing time for the network node.
  • The foregoing and other objects, features, and advantages of the claimed invention will become more apparent from the following detailed description, which proceeds with reference to the accompanying figures.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 illustrates a data transaction organized into a chain (TD=4).
  • FIG. 2 illustrates a data frame assembled by multiple basic payload frames (BPFs).
  • FIG. 3 illustrates integration of the proposed protocol in the current layered architecture.
  • FIG. 4 illustrates a simulation topology embodiment.
  • FIG. 5 is a graph of normalized goodput for BFP over Ethernet.
  • FIG. 6 is a graph of normalized goodput for TCP.
  • FIG. 7 is a graph of normalized goodput for transactions≧100 MB.
  • FIG. 8 is a graph of delay per transaction for transactions≧100 MB.
  • FIG. 9 is a graph of average buffer size for transactions≧100 MB.
  • FIG. 10 is a schematic diagram of a representative optical transport network (OTN) switch modified to support a big file protocol (BFP).
  • FIG. 11 is a schematic diagram of another representative optical transport network (OTN) switch modified to support a big file protocol (BFP).
  • DETAILED DESCRIPTION I. Explanation of Terms
  • As used in this application and in the claims, the singular forms “a,” “an,” and “the” include the plural forms unless the context clearly dictates otherwise. Additionally, the term “includes” means “comprises.” Further, the term “coupled” does not exclude the presence of intermediate elements between the coupled items.
  • The systems, apparatus, and methods described herein should not be construed as limiting in any way. Instead, the present disclosure is directed toward all novel and non-obvious features and aspects of the various disclosed embodiments, alone and in various combinations and sub-combinations with one another. The disclosed systems, methods, and apparatus are not limited to any specific aspect or feature or combinations thereof, nor do the disclosed systems, methods, and apparatus require that any one or more specific advantages be present or problems be solved. Any theories of operation are to facilitate explanation, but the disclosed systems, methods, and apparatus are not limited to such theories of operation.
  • Although the operations of some of the disclosed methods are described in a particular, sequential order for convenient presentation, it should be understood that this manner of description encompasses rearrangement, unless a particular ordering is required by specific language set forth below. For example, operations described sequentially may in some cases be rearranged or performed concurrently. Moreover, for the sake of simplicity, the attached figures may not show the various ways in which the disclosed systems, methods, and apparatus can be used in conjunction with other systems, methods, and apparatus. Additionally, the description sometimes uses terms like “produce” and “provide” to describe the disclosed methods. These terms are high-level abstractions of the actual operations that are performed. The actual operations that correspond to these terms will vary depending on the particular implementation and are readily discernible by one of ordinary skill in the art.
  • In some examples, values, procedures, or apparatus' are referred to as “lowest”, “best”, “minimum,” or the like. It will be appreciated that such descriptions are intended to indicate that a selection among many used functional alternatives can be made, and such selections need not be better, smaller, or otherwise preferable to other selections.
  • II. Introduction
  • Bandwidth demand is increasingly dominated by large transactions. The present disclosure concerns embodiments of a method and system useful for a networking approach wherein large file transactions are handled separately from the large number of smaller transactions. Traffic shaping can then be used to improve link utilization and reduce the computational load placed on the network. Traffic shaping for the corresponding long hold times can benefit from a far less dynamic traffic mix. In what follows, large transactions are shaped into periodic concatenations of data frames which are scheduled for transmission over an end-to-end path using a lightweight scheduling protocol. This Big File Protocol (BFP) shapes each transaction prior to transmission and executes the scheduling procedure on a per-flow basis. BFP creates a tighter linkage (than TCP/UDP) between the application layer and resource availability at the physical layer (either Ethernet or OTN), made possible by the long push times of long file transfers. This approach does not require network-wide coordination. Also, the disclosed methods and systems can be implemented over any packet switched network, and are entirely compatible with the current OTN infrastructure as well as with Ethernet-based networks, which are commonly used in data-center environments. The disclosed methods and systems are able to overcome the limitations of traffic shaping on end-to-end performance of the network while allowing a more efficient utilization of network resources, and a significant increase in goodput (>40%), potentially reducing the cost per bit.
  • III. Methods and Systems
  • The big file protocol disclosed herein is a cross-layer transport protocol designed to handle large transactions (e.g. 100 MB) in a more efficient manner by shaping traffic and leveraging currently available transport layer functionalities to enable a better use of the deployed capacity and reduce the computational load placed on the network by large transactions. The nodes implementing BFP shape traffic from each transaction into periodic, semitransparent concatenations of fixed-size data frames, referred to as chains henceforth.
  • Periodicity of each chain is obtained by modeling access to the transport resource as a periodic succession of fixed length timeslots in which a transaction is allowed access to an output channel at full capacity. Periodicity simplifies scheduling and stabilizes traffic variations resulting in stable network performance. The semi-transparency of each chain, resulting from periodic access to the channel, allows interleaving several flows using buffers to align the flows in time. With reference to FIG. 1, the transparency degree (TD) of a chain is defined as the period (expressed in number of timeslots) of the occupied slots in a periodic frame structure that represents capacity on a transport resource.
  • Note that TD can be dynamically adapted based on the available BW, thus a cross-layer approach may be utilized. Furthermore, BFP differs from TDM systems in that timeslots are only used to model transport resources and regulate access on a per flow basis, and need not to be physically implemented in the underlying hardware. Another difference with TDM systems is that the timeslots in TDM are somehow rigid and each TDM channel occupies a fixed amount of the available bandwidth. In BFP the amount of bandwidth occupied is decided on a per-transaction basis, resulting in a more flexible bandwidth usage. Lastly, the size of the BFP timeslots can be adjusted by varying the size of the data frames. In each chain, the fixed-size data frames are comprised of an integer number of Basic Payload Frames (BPF) assembled back to back as shown in FIG. 2. Once a receiving node delineates a data frame consisting of one or more BPFs, it handles it as a single frame. This allows for the use of data frames of virtually any size.
  • Basic Void Frames (BVF) are also defined. BVF are equal in size to BPFs but carry only stuffing bits and are added at transmission time to fill the gaps between data frames according to the selected TD. Using BPF and BVF effectively allows setting up a TDM-like channel on the fly, targeted to the specific requirements of each transaction, which is automatically released when the transaction is completed.
  • Scheduling is the means by which contending requests are directed to transmit so as to minimize buffering required to efficiently pack the transport resource. Admission control is the process used to grant or deny permission to contend.
  • A network of fixed capacity links is presumed to have been provisioned interconnecting widely separated data centers and this capacity is accessed only by BFP. This deployed capacity may be in the form of dedicated transport capacity, whether this is Ethernet, Ethernet over OTN, or a mapping of BFP directly onto the OTN layer. The data frames fit naturally within the underlying transport capacity.
  • Since BFP exploits a tight linkage between application and physical-layer resources, it can be described with reference to FIG. 3. Application programmers must see this as yet another variant of well-known socket-related actions, like TCP and UDP, as discussed later. But whereas TCP and UDP pass through IP before transport, BFP passes through a parallel path to transport. This path manages admission control, scheduling, and mapping into physical-layer resources, functions that may exploit information passed between nodes by TCP/IP.
  • IV. Admission Control and Scheduling
  • When a source is ready to transmit a chain, this is associated with a control packet (CP), which is sent over the data path whose propagation time is assumed to be known by the source node to reserve resources such as bandwidth and buffer space, using a two-way reservation protocol. Due to the periodic configuration of the data frames in a chain, a small CP can convey full information on chain configuration and timing, regardless of the amount of data transported by the chain. During the resource reservation phase the data sits at the source machine until confirmation (ACK) of successful reservation is received, avoiding unnecessary occupation of network resources. Many alternative approaches to scheduling are possible. A representative approach is described below that can be used to schedule chained data structures over an end-to-end path.
  • When a CP for the selected configuration (i.e. frame size, TD and number of data frames per file) is generated at the source node an Expected Time of Arrival (ETAsrc) parameter is computed according to the following Equation (1) and included in the CP.

  • ETAsrct=1 Nτij=1 N p jack  (1)
  • where:
    N number of nodes in the path
    τi propagation time to reach node i
    pj estimated CP processing time for node i
    τack time for the ACK to reach the source node
    The ETA parameter communicates to each node in the path the delay between the reception of the CP and the arrival of the first bit of its relative chain. Each node uses ETA and chain configuration information to compute the buffering time (BT) to interleave an incoming chain with previously scheduled chains on the output channel, according to Equation 2. Before being forwarded to the next node, CP is updated with a new ETA information, computed according to Equation 3. This updated ETA value will be used by the next node in the path to schedule the chain.

  • BTi =t e,1−ETAi  (2)
  • where:
    te,1 ending time of the first available timeslot on the outgoing channel.

  • ETAt*1=ETAi+BTi −p i  (3)
  • Once the CP reaches the destination node, an acknowledgement (ACK) is generated and sent over the reverse path to the source node. Upon receiving the ACK, the source node is assured that resources are available to support the transaction and starts transmitting the data frames separated by a number of void frames (BVF) consistent with the selected TD. Using this procedure, each node in the path can reserve resources only for the time necessary for the chain to traverse it, optimizing resource utilization. Buffering can be used to align frames in time and to overcome any timing uncertainties. Therefore, timing precision required by the scheduling procedure should be manageable using standard ranging procedures.
  • V. Mapping onto Transport Layer
  • Also described herein is the mapping of the proposed protocol onto the transport layer after a brief review of Ethernet and OTN framing.
  • A. Mapping BFP onto Ethernet
  • Standard Ethernet frames bare a payload varying from a minimum of 42 bytes to 1500 bytes. This limitation is removed when jumbo frames are used, allowing up to 9000 bytes of payload. Each Ethernet frame has a preamble of 7 bytes, 1 byte of Start Frame Delimiter (SFD), 6 bytes for source MAC address and 6 for destination MAC address, a 4 byte Frame Check Sequence is also appended to the Ethernet frame. Furthermore, after each frame is sent, transmitters are should transmit a minimum of 12 bytes of Inter Frame Gap (IFG).
  • BPF and BVF can be mapped directly onto Ethernet frames by setting the size of BPF and BVF equal to the Ethernet frame payload size (e.g.: 9000 B), allowing the proposed protocol to be mapped onto an Ethernet transport layer. In this case, frame delineation is performed by the Ethernet frames and preamble or Inter Frame Gap (IFG) bits can be used to code frame-related information (e.g. frame number, frame type—BPF, BVF or CP— and data frame size). Note that in the 10GE case, only full-duplex mode is allowed, making the preamble bits unnecessary (10GE receivers ignore preamble bits) and allowing the use of these bits for other purposes, e.g. as proprietary OAM channels.
  • B. Mapping BFP onto OTN
  • ITU-T G.709 (OTN) provides bit and timing transparent transport services. Standard containers for any client signal available today are defined together with their relative mapping procedures. Furthermore, flexible containers (i.e. ODUflex) are defined in the G.709 transport hierarchy to support packet-based clients with a wide range of bitrates. Each OTN frame bares a payload of 15232 bytes over which one or more client signals are mapped. For cell or packet based clients OTN uses GFP to encapsulate data packets and generate a continuous stream of GFP frames which is then mapped in an octet-aligned manner directly onto the OPU payload area. In this case, rate adaptation is done using GFP idle frames, which are transmitted anytime there is no data to send.
  • For implementation of BFP over the OTN transport infrastructure, the general idea is to reserve an ODU channel over each link of a network to build an overlay network which is then used exclusively to handle BFP transactions. In some embodiments, the portion of dedicated bandwidth is flexible. To achieve this flexibility without using cumbersome control plane operations, setting up ODUflex channels for each link of the overlay network seems the best option as these channels can be resized without tearing down and re-establishing the connection via the Hitless Adjustment of ODUflex protocol. The configuration of the overlay network (topology and capacities of each link) should be relatively stable (time-wise) and only occasional changes will be needed. Such changes can be planned offline over coarse timescales. Once the overlay has been established, each chain is built by mapping BPF and BVF onto GFP frames, and GFP Extension Header is used to code frame-related information. The resulting stream of GFP frames is then mapped onto the payload area of the ODUflex (i.e. ODUflex(GFP)) signal in an octet-aligned manner. Frame delineation is left to GFP and any eventual rate adaptation is performed using GFP idle frames.
  • C. Routing
  • Since the proposed approach provides a methodology separate from IP for end-to-end data transmission, a way to route chains from source to destination through a network can be used. Although routing and load balancing algorithms can be specifically designed for the disclosed methods, taking into account the periodic configuration of the proposed transport structure, it is also possible to simply re-use protocols that are already in place. In the following a method to reuse deployed routing protocols is discussed.
  • As the CP travels through each node an association between input port and output port traversed by the CP is built and stored locally in a table at each node. The assumption here is that both CP and data chain will go through the same ports in both the forward and backward direction. This port mapping is based on the particular routing protocol implemented at each specific node and the routing information is only accessed when the CP is processed avoiding any header lookup operations on the data frames of each chain. Data frames of each chain are easily identified using the TD information (periodicity) combined with the ETA parameter carried by the relative CP, and are simply delayed by an amount of time equal to the Buffering Time (which is the same for each frame of the same chain) before being switched to the output port indicated by the aforementioned table without further processing. In this respect the periodicity of each chain can be thought as an “embedded Forwarding Equivalence Class”’ (e-FEC) relative to all the frames in a chain. Each table entry is relative to a specific chain and is stored in the table until the chain traverses the node. The size of this table depends on both on the number of transactions each node is able to handle and on the average duration of each reservation procedure. Although the size of the table may become cumbersome the assumption here is that large transactions are relatively infrequent and the reservation procedure is fast enough to keep the size manageable.
  • D. Application Programming Interface
  • In some embodiments, the protocol can be installed in the network stack in a way that provides a certain level of familiarity for the application developers. That is, an application developer, given the drive to use the protocol and the familiarity with the usage of network sockets, should be able to undertake the well-known and common socket related actions—such as creating, connecting, listening, sending, receiving, etc. —in a way that is not much different from the actions that are involved in programming with TCP or UDP sockets. The proposed protocol can be registered with the socket layer, can be easily accessed by just using an indicator that will point to the newly registered protocol upon socket creation, and the functionality is provided with identical function headers. In the same sense, the structural similarities will also have a positive impact on incorporating the proposed protocol with software packages that abstract the socket layer.
  • Fortunately, the installation options already exist in the Linux Kernel. The proposed BFP can be registered with the socket layer as a new transport layer protocol making it appear to the application developers and the abstraction packages as just another protocol like TCP and UDP, hence, hiding the cross-layer behavior of the proposed protocol at the lower layers (See FIG. 3 Furthermore, implementation of the proposed protocol in the form of a loadable kernel module (LKM) will ensure that the kernel patches, hence possible regressions, are avoided, and the new protocol can be unloaded upon request. This approach also ensures that the deployment is relatively easy from the network administrators' point of view.
  • E. Hardware Implementation:
  • Fortunately, the installation options already exist in the Linux Kernel. An exemplary BFP (FIG. 10) shows how BFP functionalities can be integrated into an OTN-enabled NIC card. Incoming optical signals (OTUk) are converted into electronic format and processed by the OTN processor (block 1000) before being passed to the Fabric Interface Chip (FIC, block 1002) as packets. Here BFP traffic is separated from non-BFP traffic. BFP Control Packets (CP) are processed and information about configuration and timing of their relative chains is used by the FIC to schedule buffer space and switching resources for the data chains. CP source and destination addresses are passed to the routing layer through the Routing Layer Interface (block 1004) which communicates with whichever routing protocol is available and presents the FIC with an input-port-to-output-port mapping relative to the addresses provided. This information is used by the FIC to populate a table (referred to as the port-map table), which contains the binding of input-to-output ports for each chain.
  • Data frames from each chain are routed through a buffering stage (block 1006) to be delayed by an amount of time consistent with the buffering time (BT) computed for each chain. At the output of the buffering stage, chains can be sent to their respective output ports through the switch fabric. Chains going through the same ports will naturally interleave with one another.
  • Functional blocks of the input NIC (block 1010) are described below.
  • OTN Processor, Block 1000:
  • 1. Extracts ODUk signals from the input OTUk signals
  • 2. Packetizes ODUk signals
  • 3. Passes packetized ODUk signals to the FIC
  • Fabric Interface Chip (FIC), Block 1002:
  • 1. Separates BFP traffic from standard traffic (i.e. non-BFP traffic)
  • 2. Processes Control Packets and computes Buffering Time (BT) for each chain
  • 3. Controls the BFP buffering stage (block 1006) by setting up the BT for each incoming chain in advance
  • 4. Requests routing information to the routing layer through the Routing Layer Interface (block 1004)
  • 5. Upon receiving routing information from block 1004 schedules switching resources for incoming chains
  • 6. Populates and updates the port-map table with the routing information obtained from the routing layer
  • 7. Routes incoming chain data frames towards the BFP buffer stage (block 1006)
  • 8. Performs standard FIC functions for non-BFP traffic
  • Routing Layer Interface, Block 1004:
  • 1. Interfaces with various routing protocols and provides the FIC with the information used to build the port-map table
  • BFP Buffering Stage, Block 1006:
  • 1. Delays frames of an incoming chain by the amount indicated from the BT
  • 2. Pass outgoing chain frames to the switch fabric
  • Switch Fabric, Block 1008:
  • 3. Performs switching operations for BFP traffic and non-BFP traffic
  • Output NIC, Block 1012:
  • 1. Maps incoming packets/frames onto OTUk signals and transmits them onto the optical mesh.
  • Block 1002 can also be implemented with two separate blocks as is illustrated in FIG. 11, namely: FIC (1102, FIG. 11) and Traffic Manager and Packet Processor (1100, FIG. 11), whose functionalities are as follows:
  • Traffic manager (TM)+packet processor (PP) (illustrated as block 1100 in FIG. 11): the traffic manager separates BFP traffic (control and data) from standard traffic (i.e., non-BFP traffic). CPs are processed in this block and BT is computed for each incoming chain in advance. Based on the source and destination addresses carried by the CP, this block requests routing information to the routing layer through the routing layer interface (1004, FIG. 11). Upon receiving routing information, the TM schedules switching resources and reserves buffer space (in advance) for incoming chains. The TM also populates/updates the port-map table with the newly obtained routing information. When BFP data frames (i.e., chains) arrive at the TM, these are routed to the buffer stage (1006, FIG. 11), where each data frame is delayed by an amount of time consistent with its relative BT before being routed toward the FIC (1102, FIG. 11).
  • The functionalities of the Fabric Interface Chip (FIC, 1002 of FIG. 10/1102 FIG. 11) usually involve the following: Fabric interface chip (FIC, 1002 of FIG. 10/1102 FIG. 11)—this block performs standard FIC functions (i.e., traffic adaptation and congestion control) for both BFP and non-BFP traffic.
  • VI. Examples
  • In this exemplary embodiment, the ability to efficiently handle bulk data transfers of the disclosed method was evaluated in comparison to that of TCP Westwood, a high-speed version of TCP currently implemented in the Linux kernel. All simulations were run using Omnet++ discrete event simulation tool. The two approaches are compared in terms of goodput, end-to end delay and average buffer size. A dumbbell topology (FIG. 4) was selected for this example. The propagation delay for each link is set to 1 mS giving a RTT of roughly 6 mS. Bitrate was set to 10 Gb/s for all links. With reference to FIG. 1, the size of each data frame is set to be one BPF, and the same size is selected for the voids (i.e. one void is comprised of a single BVF) for all BFP cases. Transactions varying in size from 500 kB to 1 GB were considered and each transmission repeated until statistical stability is reached.
  • To model the time interval between two consecutive transactions (for both TCP and BFP) the system was studied using both an exponential distribution with mean μ=1 mS and a generalized Pareto distribution with scale parameter σ=0.0003 and shape parameter ξ=2. The latter case was selected to test the protocols with a more realistic traffic pattern. However, results did not show significant differences between the two distributions for the performance parameters considered (due to space constraints only results for exponential inter-arrival times are reported). For the BFP over OTN case, each OTN frame carries 15222B of payload, corresponding to the payload area of the OTN frame minus the GFP overhead including core, payload header, and two bytes of the extended header which are used as a binary counter to number each frame in a chain, allowing a maximum size for the bulk data carried in a chain of roughly 1 GB. An Ethernet jumbo frame with 9000B of payload was selected as the basic payload frame for the BFP over Ethernet case. A TD of 8 was used for all BFP cases and a maximum buffer size of 12 frames (TD+4) was selected for both BFP cases. This is not a strict requirement and can be relaxed by trading buffer space for higher utilization and lower CP blocking probability. For the TCP simulation a MSS of 8960B was selected and the layer 2 MTU was set to 9000B (Ethernet jumbo frame). A RED queuing discipline was used in each router, with: queue weight qw=0.002, minimum threshold minth=5 packets, maximum threshold maxth=50 packets and maximum packet marking probability maxpb=0.02.
  • A. Goodput
  • Instead of considering link utilization (which could be misleading, especially in the TCP case) the goodput normalized to the maximum link capacity was considered to measure the amount of payload per unit of time each approach is able to correctly deliver and the amount of bandwidth used to transfer the data payload. Protocol overhead was excluded since this is not part of the payload. FIGS. 5 and 6 show the average normalized goodput for BFP over Ethernet and TCP respectively. Performance of BFP over OTN were also studied. Results show nearly identical performance of BFP over OTN and over Ethernet. Due to space constraints only the latter is shown here.
  • For transaction sizes>10 MB, it was determined that in both BFP cases goodput increases linearly up to roughly 99% (reached when the number of sources equals the TD). This is the result of scheduled interleaving of chained data which gives a goodput close to 100% with CP blocking occurring only when normalized goodput is close to its maximum. Beyond this point goodput remains above 70% for both the BFP cases and all transaction sizes considered. Note that with BFP, small transactions (≦10 MB) can achieve high link utilization due to interleaving of flows coming from different sources.
  • In the TCP case, each source tends to fill the entire bandwidth of their link to the bottleneck router which gives higher goodput with respect to the BFP case for long lived TCP flows (≧300 MB) as long as the number of sources remains limited (≦3). For shorter flows (transaction size≦100 MB) TCP either does not ramp up fast enough to fill the available bandwidth or, as the number of sources increases, TCP is prone to packet drops which trigger the TCP backoff algorithm, reducing the offered load from the TCP sources and resulting in poor link utilization. Another issue with multiple TCP sources competing for the same resources, common in data center environments, is TCP incast, which leads to throughput collapse. While this phenomenon was observed in some examples for the TCP case it did not occur in the BFP case. FIG. 7 shows a comparison of normalized goodput between BFP and TCP for transaction sizes≧100 MB. The rapid goodput collapse observed for large transactions (≧100 MB) in TCP as the number of sources increase suggest a significant advantage relative to the amount of offered load that BFP can handle with respect to TCP. BFP can accommodate over 40% more load with respect to TCP without incurring goodput collapse.
  • B. Delay
  • As a measure of the delay performance, the average time needed to successfully complete one transaction was selected, including the time needed to setup the connection and any data retransmission. A new connection is setup for each new transaction. A comparison of the end-to-end delay performance for transactions≧100 MB is shown in FIG. 8. For transactions≧500 MB and up to 2 sources TCP tends to fill the available bandwidth and achieves smaller delays. As the number of sources increase, TCP goodput collapses (see FIG. 6) and end-to-end delay rapidly increases to values up to over 30 times larger than BFP. Similar performance can be seen for all other transaction sizes (e.g. 300 MB and 100 MB), although in these cases BFP achieves better delay performances over the entire range considered. The rapid degradation of delay performance of TCP is due to the high packet drop rate occurring when multiple sources compete for the same resources as well as to the large number of acknowledgements (ACK) used by TCP. Although in the topology studied the Round Trip Time (RTT) is relatively small, the effect of ACKs transmission has a heavy influence on TCP delay performance (each ACK will add to the overall delay an amount proportional to RTT=2). Delay will further degrade in the TCP case as the network diameter increases. This would also happen in the BFP case but to a much smaller degree due to the much smaller amount of control information exchanged between source and destination.
  • The particular buffering technique used by BFP also improves delay performance with respect to TCP by reducing the queuing delays when the network is congested (See FIG. 9). Lastly, due to the periodic configuration of data frames in each BFP transaction, the variation of the end-to-end delay over the entire range considered is much smaller for BFP, resulting in more stable delay performance of BFP over TCP.
  • C. Buffer Size
  • FIG. 9 shows a comparison of the average buffer space utilized by BFP and TCP for transactions≧100 MB. Average buffer size for BFP is proportional to the BFP frame size and the selected TD, resulting in stable and predictable buffer occupancy in any load condition. The random nature of TCP packet arrival will tend to fill the buffer quickly as congestion approaches. FIG. 9 shows that for a transaction size of 1 GB, buffer size grows rapidly to values several times larger than in the BFP case when more than 2 TCP sources compete for the same bottleneck link. Similar performance is observed for other transaction sizes. Lastly, since BFP uses buffer to delay each incoming chain in order to interleave it with previously booked chains, the maximum buffer size for BFP is a function of the chains configuration rather than of the network load, and can therefore be tightly controlled.
  • The present disclosure concerns a BFP, a network protocol designed to efficiently handle large transactions over the existing network infrastructure. Using BFP, large file transactions are handled at lower layers (e.g. L1 and L2). Higher layers are accessed only when needed with routing functionalities that are only used during connection setup and not for every single data frame as in TCP. Scheduling allows BFP to accommodate network loads over 40% higher with respect to TCP without incurring goodput collapse and achieving goodput values close to 100%.
  • Stable end-to-end delay performance is also achieved, with values over 30 times smaller with respect to TCP. Buffer occupancy of BFP in congested scenarios is predictable and much smaller relative to TCP. The results presented show the potential of BFP to achieve better use of the deployed capacity, potentially resulting in significant cost savings for the network operators in both transport and data center environments.
  • In view of the many possible embodiments to which the principles of the present disclosure may be applied, it should be recognized that the illustrated embodiments are only preferred examples of the present disclosure and should not be taken as limiting the scope of the present disclosure. Rather, the scope of the present disclosure is defined by the following claims. We therefore claim all that comes within the scope and spirit of these claims.

Claims (20)

We claim:
1. A method for transporting a data file, comprising:
dividing the data file into a plurality of data frames;
concatenating the plurality of data frames into at least one chain with a predefined transparency degree;
associating the at least one chain with a control packet comprising information regarding at least a chain length of the at least one chain and the predefined transparency degree;
scheduling transmission of the at least one chain along a predetermined path of a network based on the control packet; and
transmitting each data frame along the predetermined path at a time derived from an estimated time-to-transmit parameter.
2. The method of claim 1, further comprising concatenating the plurality of data frames into the at least one chain with one or more void frames.
3. The method of claim 1, wherein dividing the data file into a plurality of data frames comprises selecting a data frame size matching that of a transport layer.
4. The method of claim 1, further comprising using a buffer having a buffer size proportional to a data frame size and a transparency degree to delay the at least one chain so as to interleave the at least one chain with a previously received chain or an incoming chain.
5. The method of claim 4, wherein a maximum buffer size is selected based on a maximum data frame size and a maximum transparency degree of the at least one chain.
6. The method of claim 1, wherein the data file has a size of at least 100 MB.
7. The method of claim 1, wherein the predefined transparency degree of the at least one chain is associated with a number of periods of the at least one chain having at least one timeslot occupied by at least one data frame.
8. The method of claim 1, wherein the predefined transparency degree is associated with a number of void frames in the at least one chain.
9. The method of claim 1, wherein the estimated time-to-transmit parameter is included in the control packet and is communicated to at least one node of the predetermined path and is associated with a delay between control packet arrival and chain reception commencement.
10. A system for transporting a data file, comprising at least one source server comprising a software module configured to:
divide the data file into a plurality of data frames;
concatenate the plurality of data frames into at least one chain with a predefined transparency degree; and
associate the at least one chain with a control packet comprising information regarding at least a chain length of the at least one chain and the predefined transparency degree.
11. The system of claim 10, wherein the software module is further configured to:
schedule transmission of the at least one chain along a predetermined path of a network based on the control packet; and
transmit each data frame along the predetermined path at a time derived from an estimated time-to-transmit parameter.
12. The system of claim 10, wherein the at least one chain further comprises one or more void frames.
13. The system of claim 10, wherein the plurality of data frames comprises data frames of different sizes.
14. The system of claim 10, wherein the predefined transparency degree of the at least one chain is associated with a number of periods of the at least one chain comprising at least one timeslot occupied by at least one data frame.
15. The system of claim 10, wherein the predefined transparency degree is associated with a number of void frames in the at least one chain.
16. The system of claim 10, wherein the estimated time-to-transmit parameter is included in the control packet and is communicated to at least one node of the predetermined path and is associated with a delay between control packet arrival and chain reception commencement.
17. The system of claim 10, wherein the system includes a network interface that communicates the at least one chain and wherein the network interface is an optical transport network or an Ethernet-based network.
18. A network node, comprising a network interface configured to receive at least one chain that includes a plurality of data frames and void frames, or a portion of the at least one chain, wherein the network node transmits an updated estimated time-to-transmit parameter to a destination network.
19. The network node of claim 18, wherein the network interface is configured to receive a control packet associated with the at least one chain and configured to extract an estimated time-to-transmit parameter from the control packet.
20. The network node of claim 18, wherein the updated estimated time-to-transmit parameter is determined by adding an original estimated time-to-transmit parameter of the network node to an original buffering time of the network node and subtracting an estimated control packet processing time for the network node.
US14/628,102 2014-02-20 2015-02-20 System and method for efficient transport of large data files Active 2035-09-18 US9800509B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/628,102 US9800509B2 (en) 2014-02-20 2015-02-20 System and method for efficient transport of large data files

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201461942552P 2014-02-20 2014-02-20
US14/628,102 US9800509B2 (en) 2014-02-20 2015-02-20 System and method for efficient transport of large data files

Publications (2)

Publication Number Publication Date
US20150236957A1 true US20150236957A1 (en) 2015-08-20
US9800509B2 US9800509B2 (en) 2017-10-24

Family

ID=53799140

Family Applications (1)

Application Number Title Priority Date Filing Date
US14/628,102 Active 2035-09-18 US9800509B2 (en) 2014-02-20 2015-02-20 System and method for efficient transport of large data files

Country Status (1)

Country Link
US (1) US9800509B2 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017063457A1 (en) * 2015-10-12 2017-04-20 深圳市中兴微电子技术有限公司 Rate adaptation method and apparatus, and computer storage medium
US10462471B2 (en) * 2015-06-08 2019-10-29 Huawei Technologies Co., Ltd. Data processing method, communications device, and communications system
CN111200821A (en) * 2018-11-16 2020-05-26 华为技术有限公司 Capacity planning method and device
US20210226936A1 (en) * 2020-01-21 2021-07-22 The Boeing Company Authenticating computing devices based on a dynamic port punching sequence

Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4571671A (en) * 1983-05-13 1986-02-18 International Business Machines Corporation Data processor having multiple-buffer adapter between a system channel and an input/output bus
US6044061A (en) * 1998-03-10 2000-03-28 Cabletron Systems, Inc. Method and apparatus for fair and efficient scheduling of variable-size data packets in an input-buffered multipoint switch
US6233226B1 (en) * 1998-12-14 2001-05-15 Verizon Laboratories Inc. System and method for analyzing and transmitting video over a switched network
US20030016630A1 (en) * 2001-06-14 2003-01-23 Microsoft Corporation Method and system for providing adaptive bandwidth control for real-time communication
US20040190459A1 (en) * 2002-11-28 2004-09-30 Satoshi Ueda Communication apparatus, transmission apparatus and reception apparatus
US20050232150A1 (en) * 2003-06-03 2005-10-20 Kazuto Nishimura Flow control method and apparatus thereof
US7006438B2 (en) * 2001-05-31 2006-02-28 Turin Networks Distributed control of data flow in a network switch
US7035220B1 (en) * 2001-10-22 2006-04-25 Intel Corporation Technique for providing end-to-end congestion control with no feedback from a lossless network
US7039013B2 (en) * 2001-12-31 2006-05-02 Nokia Corporation Packet flow control method and device
US20070002823A1 (en) * 2005-05-09 2007-01-04 Skov Andersen Niels P Data transmission with efficient slot and block formats in a wireless communication system
US7298728B2 (en) * 2002-12-23 2007-11-20 Alcatel Lucent Scheduling system and method for a burst switch
US20090207859A1 (en) * 2003-05-14 2009-08-20 Beshai Maged E Flow-Rate-Regulated Burst Switches
US20120155476A1 (en) * 2009-06-23 2012-06-21 Roberto Pavani Clock recovery in a system which transports tdm data over a packet-switched network
US20130188568A1 (en) * 2012-01-20 2013-07-25 Chun-Ming Kuo Wireless communication apparatuses and related methods
US20130201826A1 (en) * 2010-07-09 2013-08-08 Telefonaktiebolaget L M Ericsson (Publ) Switching node with load balancing of bursts of packets
US8542585B2 (en) * 2003-12-19 2013-09-24 Broadcom Corporation Method and system for transmit scheduling for multi-layer network interface controller (NIC) operation
US9419902B1 (en) * 2013-12-30 2016-08-16 Google Inc. Method and system for network micro flow control

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9241049B2 (en) 2011-04-27 2016-01-19 Thomas E. Darcie System and method for efficient networking for large file transactions

Patent Citations (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4571671A (en) * 1983-05-13 1986-02-18 International Business Machines Corporation Data processor having multiple-buffer adapter between a system channel and an input/output bus
US6044061A (en) * 1998-03-10 2000-03-28 Cabletron Systems, Inc. Method and apparatus for fair and efficient scheduling of variable-size data packets in an input-buffered multipoint switch
US6233226B1 (en) * 1998-12-14 2001-05-15 Verizon Laboratories Inc. System and method for analyzing and transmitting video over a switched network
US7006438B2 (en) * 2001-05-31 2006-02-28 Turin Networks Distributed control of data flow in a network switch
US20030016630A1 (en) * 2001-06-14 2003-01-23 Microsoft Corporation Method and system for providing adaptive bandwidth control for real-time communication
US7035220B1 (en) * 2001-10-22 2006-04-25 Intel Corporation Technique for providing end-to-end congestion control with no feedback from a lossless network
US7039013B2 (en) * 2001-12-31 2006-05-02 Nokia Corporation Packet flow control method and device
US20040190459A1 (en) * 2002-11-28 2004-09-30 Satoshi Ueda Communication apparatus, transmission apparatus and reception apparatus
US7298728B2 (en) * 2002-12-23 2007-11-20 Alcatel Lucent Scheduling system and method for a burst switch
US20090207859A1 (en) * 2003-05-14 2009-08-20 Beshai Maged E Flow-Rate-Regulated Burst Switches
US20050232150A1 (en) * 2003-06-03 2005-10-20 Kazuto Nishimura Flow control method and apparatus thereof
US8542585B2 (en) * 2003-12-19 2013-09-24 Broadcom Corporation Method and system for transmit scheduling for multi-layer network interface controller (NIC) operation
US20070002823A1 (en) * 2005-05-09 2007-01-04 Skov Andersen Niels P Data transmission with efficient slot and block formats in a wireless communication system
US20120155476A1 (en) * 2009-06-23 2012-06-21 Roberto Pavani Clock recovery in a system which transports tdm data over a packet-switched network
US20130201826A1 (en) * 2010-07-09 2013-08-08 Telefonaktiebolaget L M Ericsson (Publ) Switching node with load balancing of bursts of packets
US20130188568A1 (en) * 2012-01-20 2013-07-25 Chun-Ming Kuo Wireless communication apparatuses and related methods
US9419902B1 (en) * 2013-12-30 2016-08-16 Google Inc. Method and system for network micro flow control

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10462471B2 (en) * 2015-06-08 2019-10-29 Huawei Technologies Co., Ltd. Data processing method, communications device, and communications system
WO2017063457A1 (en) * 2015-10-12 2017-04-20 深圳市中兴微电子技术有限公司 Rate adaptation method and apparatus, and computer storage medium
CN111200821A (en) * 2018-11-16 2020-05-26 华为技术有限公司 Capacity planning method and device
US11736954B2 (en) 2018-11-16 2023-08-22 Huawei Technologies Co., Ltd. Capacity planning method and apparatus
US20210226936A1 (en) * 2020-01-21 2021-07-22 The Boeing Company Authenticating computing devices based on a dynamic port punching sequence
US11876790B2 (en) * 2020-01-21 2024-01-16 The Boeing Company Authenticating computing devices based on a dynamic port punching sequence

Also Published As

Publication number Publication date
US9800509B2 (en) 2017-10-24

Similar Documents

Publication Publication Date Title
US11916782B2 (en) System and method for facilitating global fairness in a network
JP5706447B2 (en) Apparatus and method for synchronous network
US9967638B2 (en) Optical switching
CN100493035C (en) Method and system for improved network performance by dynamically setting a reassembly timer
US11722407B2 (en) Packet processing method and apparatus
US20230090712A1 (en) Link aggregation with data segment fragmentation
US20060268692A1 (en) Transmission of electronic packets of information of varying priorities over network transports while accounting for transmission delays
CN104982006A (en) System and method for providing a software defined protocol stack
US9800509B2 (en) System and method for efficient transport of large data files
US10728134B2 (en) Methods, systems, and computer readable media for measuring delivery latency in a frame-preemption-capable network
US11641324B2 (en) OTN transport over a leaf/spine packet network
US10015101B2 (en) Per queue per service buffering capability within a shaping window
CN109921972B (en) Method for transmitting and/or receiving data packets
WO2018107432A1 (en) Real-time video transmission method of multipath network
US20110158093A1 (en) Bandwidth control method and bandwidth control device in node device
CN113438182A (en) Flow control system and flow control method based on credit
CN115632925B (en) Time-certainty fusion network architecture and data communication method
CN101729373B (en) Method and device for dispatching multipriority traffic
EP2547053B1 (en) Distribution of data units among interfaces of a node of a packet-switched communication network
Albanese et al. Big file protocol (BFP): A traffic shaping approach for efficient transport of large files
US8515283B2 (en) Transparent fiber channel link management for protocol transport
US7292534B2 (en) Method and device for providing a minimum congestion flow of ethernet traffic transported over a SDH/SONET network
Mazzeo et al. Performance characteristics for delivering Ethernet private line service in a multi-vendor SONET environment

Legal Events

Date Code Title Description
AS Assignment

Owner name: UVIC INDUSTRY PARTNERSHIPS INC., CANADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DARCIE, THOMAS E.;ALBANESE, ILIJC;GANTI, SUDHAKAR;AND OTHERS;SIGNING DATES FROM 20150216 TO 20150317;REEL/FRAME:043606/0574

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment: 4