WO2015069584A1 - Geo-distributed disaster recovery for interactive cloud applications - Google Patents

Geo-distributed disaster recovery for interactive cloud applications Download PDF

Info

Publication number
WO2015069584A1
WO2015069584A1 PCT/US2014/063614 US2014063614W WO2015069584A1 WO 2015069584 A1 WO2015069584 A1 WO 2015069584A1 US 2014063614 W US2014063614 W US 2014063614W WO 2015069584 A1 WO2015069584 A1 WO 2015069584A1
Authority
WO
WIPO (PCT)
Prior art keywords
application
session state
state data
session
data
Prior art date
Application number
PCT/US2014/063614
Other languages
French (fr)
Inventor
Aman Kansal
Sriram Govindan
Original Assignee
Microsoft Technology Licensing, Llc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Microsoft Technology Licensing, Llc filed Critical Microsoft Technology Licensing, Llc
Priority to EP14805405.9A priority Critical patent/EP3069496B1/en
Priority to CN201480061769.2A priority patent/CN105765949B/en
Priority to BR112016010037A priority patent/BR112016010037A8/en
Publication of WO2015069584A1 publication Critical patent/WO2015069584A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • H04L67/1031Controlling of the operation of servers by a load balancer, e.g. adding or removing servers that serve requests
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/202Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant
    • G06F11/2023Failover techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/2053Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where persistent mass storage functionality or persistent mass storage control functionality is redundant
    • G06F11/2094Redundant storage or storage space
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/2097Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements maintaining the standby controller/processing unit updated
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/30Monitoring
    • G06F11/34Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment
    • G06F11/3409Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment for performance assessment
    • G06F11/3419Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment for performance assessment by assessing time
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/06Management of faults, events, alarms or notifications
    • H04L41/0654Management of faults, events, alarms or notifications using network fault recovery
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1095Replication or mirroring of data, e.g. scheduling or transport for data synchronisation between network nodes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1097Protocols in which an application is distributed across nodes in the network for distributed storage of data in networks, e.g. transport arrangements for network file system [NFS], storage area networks [SAN] or network attached storage [NAS]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/14Session management
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/40Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass for recovering from a failure of a protocol instance or entity, e.g. service redundancy protocols, protocol state redundancy or protocol service redirection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/30Monitoring
    • G06F11/3003Monitoring arrangements specially adapted to the computing system or computing system component being monitored
    • G06F11/3006Monitoring arrangements specially adapted to the computing system or computing system component being monitored where the computing system is distributed, e.g. networked systems, clusters, multiprocessor systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/06Management of faults, events, alarms or notifications
    • H04L41/0654Management of faults, events, alarms or notifications using network fault recovery
    • H04L41/0668Management of faults, events, alarms or notifications using network fault recovery by dynamic selection of recovery network elements, e.g. replacement by the most appropriate element after failure
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0852Delays
    • H04L43/0864Round trip delays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L63/00Network architectures or network communication protocols for network security
    • H04L63/04Network architectures or network communication protocols for network security for providing a confidential data exchange among entities communicating through data packet networks
    • H04L63/0428Network architectures or network communication protocols for network security for providing a confidential data exchange among entities communicating through data packet networks wherein the data content is protected, e.g. by encrypting or encapsulating the payload
    • H04L63/045Network architectures or network communication protocols for network security for providing a confidential data exchange among entities communicating through data packet networks wherein the data content is protected, e.g. by encrypting or encapsulating the payload wherein the sending and receiving network entities apply hybrid encryption, i.e. combination of symmetric and asymmetric encryption

Definitions

  • the Internet is a global data communications system that serves billions of users worldwide.
  • the Internet provides users access to a vast array of online information resources and services, including those provided by the World Wide Web, intranet-based enterprises, and the like. Thanks to the ubiquity of the various types of personal computing devices that exist today (such as personal computers, laptop/notebook computers, smart phones and tablet computers) and the ubiquity of the Internet, an evergrowing number of users in an increasing number of geographic regions routinely use an ever-growing number of web applications in their everyday lives. For examples, users rely on web applications for emailing, texting, video conferencing, listening to music, watching movies, searching for online information, storing and accessing online information, and online shopping and banking. Web applications are commonly hosted on data centers.
  • Geo-distributed disaster recovery technique embodiments described herein are generally applicable to ensuring the availability of an application that is being hosted on a current data center.
  • the geo-distributed disaster recovery technique embodiments provide disaster recovery for the application.
  • an option for replicating session state data for the application is selected, where this selection is made from a set of different session state data replication options each of which has different performance and resource cost trade-offs, and the selected option determines how the session state data for the application is to be replicated.
  • the selected option is then implemented, where the implementation results in the session state data for the application being replicated outside of the current data center, thus ensuring that this data remains available in the event that the current data center goes offline.
  • a client computer receives a message from the current data center specifying an option for replicating session state data for the application that has been selected for the client computer, where this selection has been made from a set of different session state data replication options each of which has different performance and resource cost trade-offs, and the specified option determines how the session state data for the application is to be replicated.
  • the specified option is then implemented, where the implementation results in the session state data for the application being replicated outside of the current data center, thus ensuring that this data remains available to the client computer in the event that the current data center goes offline.
  • Yet another exemplary embodiment involves provisioning and using spare data center capacity in a network of data centers.
  • a determination is made as to an amount of spare server capacity that is to be added to each of the data centers in the network of data centers in order for this network to be able to take over the current application load of any one of the data centers if it fails.
  • This determination is made by taking into account different factors that include one or more of: the total number of data centers, or the current server capacity of each of the data centers, or the geographic location of each of the data centers. It is then directed that the determined amount of spare server capacity be added to each of the data centers. Then, whenever any one of the data centers fails, the current application load of the failed data center is distributed across the spare server capacity that was added to each of the remaining data centers.
  • FIG. 1 is a diagram illustrating an exemplary embodiment, in simplified form, of an architectural framework for implementing the geo-distributed disaster recovery technique embodiments described herein.
  • FIG. 2 is a table summarizing different session state data replication options that can be used by the geo-distributed disaster recovery technique embodiments described herein and the performance and resource cost parameters that are associated with each of the replication options.
  • FIG. 3 is a flow diagram illustrating one embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for a cloud application that is being hosted on a data center.
  • FIG. 4 is a flow diagram illustrating another embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application.
  • FIG. 5 is a flow diagram illustrating yet another embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application.
  • FIG. 6 is a flow diagram illustrating an exemplary embodiment, in simplified form, of the server side of a process for ensuring the availability of the cloud application.
  • FIG. 7 is a flow diagram illustrating an exemplary embodiment, in simplified form, of the client side of a process for providing disaster recovery for the cloud application.
  • FIG. 8 is a flow diagram illustrating an exemplary embodiment, in simplified form, of a process for provisioning and using spare data center capacity in the cloud.
  • FIG. 9 is a diagram illustrating a simplified example of a general-purpose computer system on which various embodiments and elements of the geo-distributed disaster recovery technique, as described herein, may be implemented.
  • references herein to "one embodiment”, or “another embodiment”, or an “exemplary embodiment”, or an “alternate embodiment”, or “one implementation”, or “another implementation”, or an “exemplary implementation”, or an “alternate implementation” means that a particular feature, a particular structure, or particular characteristics described in connection with the embodiment or implementation can be included in at least one embodiment of the disaster recovery technique.
  • cloud application is used herein to refer to a web application that operates in the cloud and can be hosted on (e.g., deployed at) a plurality of data centers that can be located in different geographic regions (e.g., different regions of the world) and can be concurrently used by a plurality of remote end users (hereafter simply referred to as users).
  • a cloud application is a web application that is geographically distributed.
  • client computing device is used herein to refer to a computing device that a user uses to establish an interactive session with a given a given interactive cloud application.
  • the disaster recovery technique embodiments described herein provide geo-distributed disaster recovery for such applications, while at the same time minimizing the performance costs (e.g., minimizing client latencies) and minimizing the resource costs (e.g., network bandwidth costs and data storage costs, among others) associated with providing this disaster recovery.
  • the disaster recovery technique embodiments generally involve ensuring (e.g., maximizing) the availability of interactive cloud applications so that the applications and their most recent (e.g., freshest) state data remain available to users even if one or more of the data centers that host the applications go offline for whatever reason.
  • Session state data is associated with an interactive session that one or more users establish with the application. It will be appreciated that the session state can change incrementally each time a user interacts with the application.
  • Persistent state data has two properties that the session state data does not have namely durability and concurrency.
  • the durability property of persistent state data refers to the fact that the persistent state data persists after all of the users have left the interactive session, and after the interactive session ends. In contrast, the session state data persists just for the life of the interactive session, and can be discarded once all of the users have left the interactive session.
  • the concurrency property of persistent state data refers to the fact that a plurality of users and a plurality of interactive sessions can use the persistent state data at the same time. Many of the user's interactions with the application affect the session state data but do not affect the persistent state data.
  • all of the disaster recovery technique embodiments described herein focus specifically on replicating the session state data for an interactive session that takes place over a network between a given interactive cloud application and one or more client computing devices.
  • some of the disaster recovery technique embodiments can also be used to replicate persistent state data for the application.
  • the disaster recovery technique embodiments ensure that the most recent session state data will remain available to the client computing devices in the event that the data center that is currently hosting the application and serving the session goes offline and thus becomes unable to continue serving the session.
  • various techniques exist for protecting the persistent state data in the event of a data center failure one example of which is the conventional Paxos family of protocols.
  • the fact that the disaster recovery technique embodiments treat the session state data separately from the persistent state data does not require significant modifications to existing web applications because such applications already preserve the session state explicitly between successive client requests by using a session state application programming interface (API) that is provided by the development platform for the web applications.
  • the disaster recovery technique embodiments maintain consistency between the session state data and persistent state data since the embodiments store and will recover the most recent session state data whenever the data center that is currently serving the session goes offline for whatever reason, and thus becomes unable to continue serving the session.
  • the disaster recovery technique embodiments allow developers of a given interactive cloud application to ensure the availability of the application to users by replicating the session state data for the application. As such, the session state data for interactive sessions with the application is protected against data center failures. As will be appreciated from the more detailed description that follows, the disaster recovery technique embodiments replicate the entire session state, including both user-entered data and server-computed data.
  • the disaster recovery technique embodiments provide a set of different options for replicating the session state data for the application, where each of the different session state data replication options has different performance and resource cost trade-offs.
  • the disaster recovery technique embodiments also provide various methods for automatically selecting an appropriate one of the session state data replication options that increases (e.g., maximizes) the performance of a given interactive session with the application (and hence the performance that the users of the application experience) and/or decreases (e.g., minimizes) the resource costs (e.g., decreases the resources that are utilized) for the interactive session. This automatic selection is based on various performance constraints and resource cost constraints, and a prioritization thereof, that can be specified by the application.
  • the disaster recovery technique embodiments can automatically switch from a previously selected session state data replication option to a newly selected session state data replication option in the middle of the interactive session based on changes in these constraints, or changes in their prioritization, or changes in prescribed run-time conditions in the computing and networking infrastructure that supports the application.
  • these run-time conditions can include, but are not limited to, various performance and resource costs such as network bandwidth costs, network latency experienced by client computing devices, data storage costs, processing costs, and battery energy drains costs, among others.
  • the run-time conditions can also include, but are not limited to, various resource availabilities such as the current availability (e.g., percent utilization) of computing device processors, memory, and batteries, among others.
  • Each of these run-time conditions can be different for different client and server computing devices in the infrastructure, and for different portions of the network in the infrastructure.
  • Each of the run-time conditions can also change over time.
  • the disaster recovery technique embodiments described herein are advantageous for various reasons including, but not limited to, the following.
  • the disaster recovery technique embodiments make interactive cloud applications resilient to data center failures by replicating the session state data for the applications.
  • the disaster recovery technique embodiments minimize the computing and network transmission costs associated with this session state data replication so that the performance of interactive sessions with the applications is maximized, the user response times for the applications are minimized, and the resource costs for the data centers that host the applications are minimized.
  • the disaster recovery technique embodiments also allow regional legal requirements to be met (e.g., current regulations in some countries dictate that some of the data that is served by a data center be located within national boundaries).
  • the disaster recovery technique embodiments also integrate easily with existing cloud applications, in some cases without source code modifications.
  • the disaster recovery technique embodiments automatically adapt to network bandwidth variations at the client computing devices (e.g., network bandwidth variations caused by either cellular or WiFi network signal strength
  • the disaster recovery technique embodiments also automatically adapt to variable data center placements and transparently benefit cloud applications as new data centers are added to the cloud.
  • the disaster recovery technique embodiments described herein are further advantageous in that they provide usage scalability for interactive cloud applications (e.g., the disaster recovery technique embodiments ensure the availability of a given interactive cloud application even if a large number of users is using the application at the same time).
  • the disaster recovery technique embodiments described herein also allow interactive sessions that take place over a network between interactive cloud applications and client computing devices to be served by nearby data centers.
  • the disaster recovery technique embodiments ensure that the data center that is hosting a given interactive cloud application and currently serving a given interactive session between the application and a given client computing device will be close to the client computing device regardless of which geographic region the client computing device is in. This close proximity of the data center and client computing device reduces client latency and user response times.
  • a user of a given interactive cloud application is particularly sensitive to the performance of the application during the interactive phases of a session with the application.
  • a user of an ecommerce website may want fast response time when browsing for products on the website, but may accept slower response times during the final stage of making a product purchase involving credit card verification.
  • the disaster recovery technique embodiments described herein assume that an interactive session with the application has begun.
  • Data that is relevant to the interactive session can be routinely generated by both the user and the application as the session progresses.
  • the website may have generated data about which available flights operate between the desired departure and destination cities on the desired dates.
  • the user may have also entered filters for the flight search (e.g., the user may have requested to see just available flights that depart after 5:00 PM and have at most one stop, among other types of filters).
  • the session state data would include both the data that was entered by the user (e.g., the desired departure and destination cities, the desired dates, and the filters) and the data that was generated by the website in response to this user-entered data (e.g., which available flights operate between the desired departure and destination cities on the desired dates and comply with the filters).
  • the disaster recovery technique embodiments described herein ensure that the most recent session state data remains available to a given interactive cloud application even if the data center that is hosting the application and currently serving an interactive session over a network between the application and one or more client computing devices goes offline for any reason and another data center that is also hosting the application takes over serving the session.
  • the disaster recovery technique embodiments ensure that the other data center that takes over serving the session has access to the most recent session state data.
  • the user will not have to re-enter any of data that they previously entered, and the other data center will not have to re-compute any of the data that was previously computed in response to the user-entered data. If the user then enters another filter that requests to see just the available flights that operate through a particular connecting airport, this other filter will be routed to the other data center, which will use the most recent session state data to generate new data for the user indicating which of the previously selected and filtered flights comply with this other filter.
  • FIG. 1 illustrates an exemplary embodiment, in simplified form, of an architectural framework for implementing the disaster recovery technique embodiments described herein.
  • the architectural framework 100 includes one or more client computing devices 102/116, a network 108, and the cloud 138.
  • the cloud 138 generally includes a plurality of data centers (e.g., 104/106/118) that can be located in different geographic regions.
  • the network 108 allows the client computing devices 102/116 to communicate with and exchange information with the data centers 104/106/118, and also allows the data centers to communicate with and exchange information with each other. It will be appreciated that the network 108 can employ a wide variety of local and wide area networking
  • each of the client computing devices 102/116 can be any type of computing device.
  • each of the client computing devices 102/116 can be either a mobile computing device (such as a smartphone, or a tablet computer, or a personal digital assistant, or the like) or a non-mobile computing device (such as a desktop personal computer (PC), or a laptop computer, or the like).
  • a mobile computing device such as a smartphone, or a tablet computer, or a personal digital assistant, or the like
  • a non-mobile computing device such as a desktop personal computer (PC), or a laptop computer, or the like.
  • users can use the client computing devices 102/116 to establish an interactive session 112 with a desired interactive cloud application 114/132 (hereafter sometimes simply referred to as a cloud application, or sometimes even more simply referred to as an application) that can be hosted on the data centers 104/106/118. More particularly, when a given user who is using a given client computing device 102 wants to use a desired cloud application 114/132, various things happen including, but not limited to, the following.
  • the application client 110 that is associated with the application 114/132 is run on the client computing device 102.
  • One of the data centers 104 that is hosting the application 114 is selected to serve an interactive session 112 between the application and the application client 110.
  • the data center that is currently selected to serve the interactive session 112 is herein referred to as the current data center 104.
  • one or more other data centers 106/118 that host a copy of the application 132 are selected to serve the interactive session 112 in the event that the current data center 104 becomes unable to do so for whatever reason. It will be appreciated that the current data center 104 can become unable to continue serving the interactive session 112 for various reasons including, but not limited to, the following.
  • the current data center 104 can go offline due to either a failure within the current data center or a failure in the network 108, or the current data can become overloaded.
  • the other data centers that are selected to serve the interactive session 112 in the event that the current data center 104 becomes unable to do so are herein referred to as backup data centers 106/118.
  • the disaster recovery technique embodiments described herein assume that the copy of the application 132 that is hosted on the backup data centers 106/118 can provide the same functionality as the application 114 that is hosted on the current data center 104.
  • each of the data centers 104/106/118 in the architectural framework 100 can include a plurality of server computing devices (not shown) which communicate over an internal network (not shown) in the data center.
  • large data centers include a large number of server computing devices
  • small data centers include a small number of server computing devices.
  • the disaster recovery technique embodiments described herein can be used in various framework 100 configurations ranging from a configuration that includes a small number of large data centers to a configuration that includes a large number of small data centers.
  • the application 114 is run on one of the server computing devices in the current data center 104.
  • the interactive session 112 that is taking place over the network 108 between the application 114 and the application client 110 is served by one of the server computing devices in the current data center 104.
  • the server computing device in the current data center 104 that runs the application 114 and serves the interactive session 112 is hereafter simply referred to as an application server.
  • the selection of which one of the data centers 104/106/118 that hosts the application 114/132 will be the current data center 104, and which one or more of these data centers will be backup data centers 106/118, can be made dynamically (e.g., on-the-fly) using various conventional methods.
  • this selection can be made by either a network of edge nodes (not shown) that is deployed for the application 114/132, or a content delivery network (also known as a content distribution network or CDN for short) (not shown), or by the application client 110 itself (as is the case when using a round robin Domain Name System).
  • the disaster recovery technique embodiments described herein assume that these selections have already been made. Once the current data center 104 and the backup data center(s) 106/118 have been selected, an interactive session 112 can be established over the network 108 between the application client 110 and the application 114 that is running in the current data center. The disaster recovery technique embodiments assume that this preselection of the current data center 104 and the backup data center(s) 106/118 is known for each interactive session 112.
  • the determination of which of the backup data centers will be selected to become the new current data center for each existing and new interactive session 112 can be made in various ways.
  • this selection can be made based on static provisioning decisions (e.g., if two backup data centers 106 and 118 are currently in operation, one of these backup data centers 106 can take over serving one half of the existing and new interactive sessions 112, and the other of these backup data centers 118 can take over serving the other half of the existing and new interactive sessions).
  • this selection can be made based on the current load conditions in each of the backup data centers 106/118 when the interactive session 112 is first established.
  • the just-described disaster recovery information exchanges can be implemented in various ways.
  • the disaster recovery information exchange can be piggybacked on (e.g., included within) a given application-related message that is to be exchanged over the network 108 between the client computing device and the current data center.
  • the disaster recovery information exchange can be implemented by the additional program modules transmitting their own messages over the network 108 which are independent of the application-related messages. It is also possible for disaster recovery information exchanges between a given client computing device (e.g., 102) and the current data center 104 to be implemented by the additional program modules transmitting their own messages over the network 108 which are independent of the application-related messages.
  • these additional program modules can be run on each of the client computing devices 102/116.
  • these additional program modules include, but are not limited to, a client state store module 120, a failure detector module 122, and a client resource probe module 124. Exemplary functions of each of these additional program modules 120/122/124 are described in more detail hereafter. As will be appreciated from the more detailed description of the different session state data replication options that also follows, some of these options necessitate the use of all of the additional program modules 120/122/124, and others of these options necessitate the use of just a subset of the additional program modules.
  • client computing devices 102/116 may be running all of the additional program modules 120/122/124 while other client computing devices may be running just a subset of the additional program modules.
  • the particular session state data replication option that is currently selected and being implemented for a given client computing device determines which of the additional program modules 120/122/124 are running on the client computing device at a given point in time.
  • the client state store module 120 stores session state data for the interactive session 112 on the client computing device 102 for the session state data replication options that involve storing the session state data on the client computing device.
  • the client state store module inputs a data object (not shown) that includes the most recent session state data and stores it on the client computing device 102.
  • the client state store module can also retrieve and output this stored data object whenever the application client 110 requests it.
  • the client state store module thus ensures that the most recent session state data remains available to the application client as long as the client computing device is available (e.g., remains functional).
  • the client state store module 120 can be implemented in various ways.
  • the client state store module can use the cookie mechanism that is offered by conventional web browsers to store data on the client computing device 102.
  • the client state store module can use a JavaScript program that stores the data object on the client computing device in either volatile memory or various types of non- volatile memory (e.g., persistent storage) that may be available on the client computing device. Examples of such non- volatile memory include, but are not limited to magnetic disk storage and flash memory.
  • the client state store module can use the conventional web storage APIs that are supported by today's web browser applications to write the most recent session state data to secondary storage such as a blob storage element in the cloud 138. This particular embodiment is advantageous in that it minimizes memory overhead on the client computing device.
  • the failure detector module 122 operates when the session state data for the interactive session 112 is being replicated using a fourth one of the different session state data replication options.
  • the failure detector module operates in two phases, namely a failure detection phase and a subsequent recovery phase.
  • the features of the failure detector module are advantageous in that they minimize the time it takes to recover from the current data center 104 going offline.
  • the disaster recovery technique embodiments described herein assume a conventional fail-stop failure model for the various data centers 104/106/118 in the cloud 138.
  • a distributed failure detector has to choose between completeness (e.g., every failure is eventually detected) and accuracy (e.g., false positives).
  • completeness e.g., every failure is eventually detected
  • accuracy e.g., false positives.
  • the disaster recovery technique embodiments described herein provide a failure detection solution that is fast and ensures completeness, but does not necessarily ensure accuracy. It is noted that alternate embodiments of the disaster recovery technique are possible that are more complex but can ensure both completeness and accuracy.
  • the failure detector module 122 detects when the current data center 104 that is serving the interactive session 112 goes offline and thus becomes unable to continue serving the session.
  • the failure detector module can be implemented in various ways. By way of example but not limitation, in one embodiment of the disaster recovery technique described herein the failure detector module can use a timer to measure the time that has elapsed since the application client 110 last submitted a client request message over the network 108 to the application 114.
  • the failure detector module can assume that the current data center has failed and gone offline (although it will be appreciated that the actual failure that resulted in the current data center's inability to respond could be somewhere in the network or elsewhere).
  • the failure detector module can periodically transmit a test message over the network to the current data center and receive a test response message therefrom that indicates the current availability thereof (e.g., this test response message may indicate that the current data center is either fully available, or is currently getting overloaded and hence may not be available in the near future, or is currently operating from battery backup power and hence may fail in the near future, among other things).
  • the failure detector module can periodically test the availability of the network at the client computing device 102 using a mechanism that is independent of the current data center, and the failure detector module can use this network availability information to enhance its knowledge about why the current data center is currently unable to serve the interactive session.
  • the recovery phase is started by the application client 110 transmitting a session resume request message over the network 108 to a backup data center (e.g., 106) that has been selected to take over serving the interactive session 112, where this message includes the most recent session state data for the interactive session that is stored on the client computing device 102.
  • a backup data center e.g. 106
  • the selected backup data center will extract the most recent session state data from the message and provide it to the copy of the application 132 so that the interactive session 112 can be resumed.
  • the application is paused until the most recent session state data has been recovered and the interactive session has been resumed at the selected backup data center. This pausing of the application can make the user think that the application has frozen.
  • the failure detector module detects that the current data center has gone offline the failure detector module can optionally inform the application client that the current data center has gone offline, and the application client can optionally display a message to the user telling them about this and that a wait is expected until the interactive session is resumed.
  • the application client can optionally display another message to the user telling them that the application is once again available for use.
  • the client resource probe module 124 operates cooperatively with a current server resource probe module 130 that is running on the current data center 104 to periodically measure relevant network performance conditions and resource availability conditions at the client computing device 102.
  • a decision engine module 126 that is also running on the current data center 104 can use these measured conditions, along with other measured conditions described herein, to select an appropriate session state data replication option. More particularly and by way of example but not limitation, in one embodiment of the disaster recovery technique described herein the client resource probe module 124 can receive server probe messages that are periodically sent to the client resource probe module by the current server resource probe module 130, and can immediately respond to each of these messages.
  • the current server resource probe module 130 can measure the current network latency associated with communicating with the client computing device 102 (e.g., the current network round trip time for the current data center 104 to reach the client computing device) simply by measuring the time between when the server resource probe module transmits each server probe message over the network 108 to the client resource probe module 124, and when the server resource probe module receives a response to the server probe message from the client resource probe module. This measurement provides an estimate of the current network latency at zero payload. It will be appreciated that since each server probe message has a very tiny payload, the network 108 overhead associated with the server probe messages and their associated responses is negligible.
  • the client resource probe module 124 can periodically measure the current network latency associated with communicating with the current data center 104 and transmit this network latency measurement to the current data center.
  • the client resource probe module can also periodically measure the current network bandwidth that is available to the client computing device 102 and transmit this network bandwidth measurement to the current data center.
  • the client resource probe module can also periodically measure the current availability of one or more resources in the client computing device (including, but not limited to, its current battery availability (e.g., percent utilization), its current processor availability, and its current memory availability) and transmit these resource availability measurements to the current data center.
  • the client resource probe module 124 can be
  • the client resource probe module can periodically measure the current network latency associated with communicating with the current data center 104, and the current network bandwidth that is available to the client computing device, simply by measuring the time between when the application client 110 transmits each client request message over the network 108 to the application 114, and when the application client receives an application response to the client request message from the application.
  • the client resource probe module can periodically transmit a client probe message over the network to the current data center, and then use a received probe response message from the current data center to determine the current network latency associated with
  • the client resource probe module can use one or more APIs that may exist in the operating system of the client computing device to determine the current availability of the just-described resources in the client computing device. It will be appreciated that there may be times when it may not be possible to measure certain resources in the client computing device (e.g., when the application client is a webpage that is running in a web browser, the application client may not be allowed to make certain resource measurements).
  • various additional program modules can be run in the current data center 104.
  • these additional program modules include, but are not limited to, a decision engine module 126, a current state store module 128, and a current server resource probe module 130. Exemplary functions of each of these additional program modules 126/128/130 are described in more detail hereafter.
  • the decision engine and current state store modules 126 and 128 can be run either on the application server (not shown), or on another server computing device in the current data center 104.
  • the current server resource probe module 130 is generally run on the application server.
  • the current server resource probe module 130 operates cooperatively with both the client resource probe module 124 and a backup server resource probe module 134 that is running on each of the backup data centers 106/118 to periodically measure relevant network performance conditions and resource availability conditions at the application server in the current data center 104.
  • the current server resource probe module 130 transmits these measurements to the decision engine module 126.
  • the decision engine module 126 can use these measured conditions, along with other measured conditions described herein, to select an appropriate session state data replication option. More particularly and by way of example but not limitation, the current server resource probe module 130 can periodically measure the current network latency associated with communicating with the client computing device 102, and the current network bandwidth that is available to the client computing device.
  • these current network latency and current network bandwidth measurements can be made by the current server resource probe module 130 periodically transmitting a server probe message over the network 108 to the client computing device 102, and then using a received probe response message from the client computing device to determine the current network latency associated with communicating with the client computing device, and the current network bandwidth that is available to the client computing device.
  • the current server resource probe module 130 can also periodically measure the current network latency associated with communicating with each of the backup data centers 106/118, and the current network bandwidth that is available to each of the backup data centers.
  • these current network latency and current network bandwidth measurements can be made by the current server resource probe module 130 periodically transmitting a server probe message over the network 108 to each of the backup data centers 106/118, and then using a received probe response message from each of the backup data centers to determine the current network latency associated with communicating with each of the backup data centers, and the current network bandwidth that is available to each of the backup data centers.
  • the current server resource probe module 130 can also periodically measure the current availability of one or more resources in the application server including, but not limited to, its current processor availability and its current memory availability. It will be appreciated that the current server resource probe module can make these resource availability measurements in various ways using various mechanisms that are available on the application server.
  • the current server resource probe module can use one or more APIs that may exist in the operating system of the application server to determine the current availability of the just-described resources in the application server.
  • the current server resource probe module can also include functionality that the client resource probe module 124 can use to measure the current network bandwidth that is available to the client computing device 102.
  • the current server resource probe module can respond to the client probe message by transmitting a data payload over the network to the client resource probe module.
  • the client resource probe module can measure the time it take to receive the data payload, and can then use this time and the size of the payload to calculate the current network bandwidth that is available to the client computing device.
  • the decision engine module 126 periodically receives information over the network 108 from the client resource probe module 124, where this information includes the most recent network performance and resource availability measurements the client resource probe module has made at the client computing device 102.
  • the decision engine module 126 also periodically receives information from the current server resource probe module 130, where this information includes the most recent performance and resource availability measurements the current server resource probe module has made at the application server in the current data center 104.
  • the decision engine module 126 also receives the aforementioned various performance constraints and resource cost constraints, and a prioritization thereof, that can be specified by the application 114.
  • Exemplary performance constraints that can be specified by the application include, but are not limited to, either a maximum allowable latency cost to the client computing device 102 associated with replicating the session state data, or a minimum allowable availability value for the application 114/132, or both this maximum allowable latency cost and this minimum allowable availability value. Exemplary resource costs and related constraints that can be specified by the application are described in more detail hereafter.
  • the decision engine module 126 selects an appropriate option for replicating the session state data for the interactive session 112 with the application 114 that increases the performance of the interactive session and/or decreases the resource costs for the interactive session.
  • this selection is made from a set of different session state data replication options each of which has different performance and resource cost tradeoffs, and can be based on the just- described constraints and their priorities.
  • the decision engine module 126 determines how the session state data for the application 114 will be stored and replicated in the framework 100.
  • the decision engine module 126 can use various methods to make this selection, examples of which are described in more detail hereafter.
  • the decision engine module can monitor the information it receives from the client and current server resource probe modules 124 and 130 on an ongoing basis.
  • the decision engine module 126 can also monitor the various performance constraints and resource cost constraints, and the prioritization thereof, that can be specified by the application 114 on an ongoing basis.
  • the decision engine module 126 can again select an appropriate one of the session state data replication options that increases the performance of the interactive session 112 with the application 114 and/or decreases the resource costs for the interactive session, where this newly selected option can be either the same as or different than the previously selected option.
  • the current state store module 128 stores the session state data for the interactive session 112 in the current data center 104 for the session state data replication options that involve storing the session state data in the current data center.
  • the current state store module 128 inputs a data object (not shown) that includes the most recent session state data and stores it in the current data center 104.
  • the current state store module 128 can also retrieve and output this stored data object whenever the application 114 requests it.
  • the current state store module 128 thus ensures that the most recent session state data remains available while the interactive session 112 is active and as long as the current data center 104 is available.
  • the current state store module 128 can store the data object in the current data center 104 in various ways.
  • the current state store module 128 can store the data object in either volatile memory or various types of non- volatile memory (e.g., magnetic disk storage, or the like) that may be available in the current data center 104.
  • the current state store module 128 can also store the data object redundantly in the current data center 104 to protect against server failures there -within.
  • each of the backup data centers 106/118 hosts a copy of the application 132.
  • one of the backup data centers e.g., 106
  • the copy of the application 132 will be run on one of the server computing devices in the selected backup data center 106 and the interactive session 112 will resume over the network 108 between this copy of the application and the application client 110. It is noted that until such a take-over occurs, the copy of the application 132 remains unused.
  • the server computing device in the selected backup data center 106 that runs the copy of the application 132 and takes over serving the interactive session 112 is hereafter simply referred to as a backup application server.
  • each of the backup data centers 106/118 can be run on each of the backup data centers 106/118.
  • these additional program modules include, but not limited to, the aforementioned backup server resource probe module 134 and a backup state store module 136. Exemplary functions of each of these additional program modules 134/136 are described in more detail hereafter.
  • the backup state store module 136 is run either on the backup application server (not shown), or on another server computing device in the selected backup data center 106.
  • the backup server resource probe module 134 is generally run on the backup application server.
  • the backup state store module 136 receives the most recent session state data for the interactive session 112 from the current data center 104 and stores this received session state data in the backup data center 106 for the session state data replication options that involve storing the session state data in the backup data center.
  • the backup state store module 136 inputs a data object (not shown) that includes the received session state data and stores it in the backup data center 106.
  • the backup state store module 136 can also retrieve and output this stored data object whenever the copy of the application 132 requests it.
  • the backup state store module 136 thus ensures that the session state data remains available while the interactive session 112 is active and as long as the backup data center 106 is available.
  • the backup state store module 136 can store the data object in the backup data center 106 in various ways.
  • the backup state store module 136 can store the data object in either volatile memory or various types of non- volatile memory that may be available in the backup data center 106.
  • the backup state store module 136 can write the data object to secondary storage such as a blob storage element which can be located either within the backup data center 106 or elsewhere in the cloud 138. This particular embodiment is advantageous in that it enhanced scalability and minimizes memory overhead.
  • the backup state store module 136 can also store the data object redundantly in the backup data center 106 to protect against server failures there-within. It will be appreciated that the backup state store module 136 can store session state data for a plurality of interactive sessions that are served by the current data center 104.
  • the backup server resource probe module 134 implements remote endpoint functionality that operates cooperatively with the current server resource probe module 130 and supports its aforementioned periodic measurement of the current network latency associated with communicating with the backup data center 106, and the current network bandwidth that is available to the backup data center 106. More particularly and by way of example but not limitation, the backup server resource probe module 134 can receive the server probe messages that are periodically sent by the current server resource probe module 130, and can immediately respond to each of these messages.
  • the current server resource probe module 130 can determine the current network latency associated with communicating with the backup data center 106 by measuring the time between when each of the server probe messages is sent and when the corresponding probe response message is received (e.g., by measuring the current round trip network delay between the current data center 104 and the backup data center 106).
  • the replication of session state data that is provided for by the disaster recovery technique embodiments described herein ensures that the session state data for the application 114/132 remains available to the application client 110 (e.g., none of the session state data is lost) in the event that the current data center 104 goes offline for whatever reason, and thus becomes unable to continue serving the interactive session 112.
  • the disaster recovery technique embodiments replicate and thus protect the entire session state, including both user-entered data and server-computed data.
  • the disaster recovery technique embodiments allow the entire state of the interactive session that is taking place between the application client and the application to be maintained, even if the current data center goes offline and one of the backup data centers 106/118 takes over serving the interactive session. As such, if the current data center were to go offline for whatever reason, none of the session state data would have to be re-entered by the user or re-computed by the backup data centers.
  • the disaster recovery technique embodiments described herein are discussed in the context of an interactive session being established between a given cloud application and a single client computing device that is being used by a single user, it is noted that the disaster recovery technique embodiments also support interactive sessions with cloud applications that include a plurality of users and a plurality of client computing devices which are jointly participating in the interactive session, and which may be located in different geographic regions. In other words, the disaster recovery technique embodiments support a situation where just a single application client is participating in an interactive session.
  • the disaster recovery technique embodiments also support a situation where a plurality of application clients running on different client computing devices is collaboratively participating in a common interactive session at the same time (examples of such a situation include a plurality of geographically distributed users collaboratively playing a multi-player online game, or collaboratively participating in an online video teleconference, or collaboratively editing an online document, or collaboratively using a given enterprise application, among others).
  • the disaster recovery technique embodiments can select different session state data replication options for different ones of the application clients that are participating in the session depending on the individual network performance conditions and resource availability conditions at the client computing devices.
  • any one or more of the application clients can leave at any time during the session, and the disaster recovery technique embodiments will ensure that the session continues correctly for the remaining application clients.
  • the disaster recovery technique embodiments are discussed herein in the context of a given cloud application serving a single interactive session, it is noted that the disaster recovery technique embodiments also support a scenario where the application is serving a plurality of interactive sessions at the same time. In such a scenario the disaster recovery technique embodiments are implemented independently for each of the interactive sessions that is being served by the application.
  • this section describes various embodiments of the disaster recovery technique described herein where the aforementioned additional program modules that can be run on the client computing devices, the current data center, and the backup data centers can be implemented as generic libraries (not shown).
  • the disaster recovery technique embodiments described in this section are hereafter simply referred to as generic libraries embodiments of the disaster recovery technique described herein.
  • the generic libraries embodiments of the disaster recovery technique automatically resume an active interactive session with a cloud application at a backup data center and recover the most recent session state data in the event that the current data center goes offline for whatever reason.
  • the generic libraries embodiments support both request-response based Hyper Text Transfer Protocol (HTTP) cloud applications and continuous interaction based web-sockets cloud applications.
  • HTTP Hyper Text Transfer Protocol
  • the generic libraries embodiments override the existing session state APIs in the cloud application and the existing communication APIs in its application client.
  • each of the generic libraries sits just below the application layer in the network stack. Accordingly, interactions with the application that would normally occur through HTTP or web-sockets will occur through the generic libraries.
  • the generic libraries intercept the HTTP and web-socket methods that applications use to communicate over the network, and add additional information (that is used to support the disaster recovery technique embodiments described herein) to application messages before they are transmitted over the network. Examples of such additional information include backup data center addresses and session state timestamps, among other types of information.
  • this additional information is added to the application messages using JSON (JavaScript Object Notation) encoding. When the messages are received on the other end of the network, the additional information is removed by the generic libraries before the messages are passed to the application.
  • JSON JavaScript Object Notation
  • the generic libraries embodiments of the disaster recovery technique are advantageous for various reasons including, but not limited to, the following.
  • the generic libraries embodiments allow the additional program modules to be easily integrated with existing cloud applications without having to make significant code modifications thereto.
  • the generic libraries embodiments also allow the cloud application developers to benefit from the functionality of the additional program modules without having to understand the details of their underlying implementation.
  • the generic libraries embodiments also allow each of the additional program modules to operate in an application independent manner such that a given additional program module can be shared by different application clients or different cloud applications.
  • the generic libraries embodiments thus eliminate the need for each different application that is hosted on the current data center to separately implement the decision engine module, the current state store module, and the current server resource probe module.
  • the generic libraries embodiments also eliminate the need for each different copy of the application that is hosted on each backup data center to separately implement the backup server resource probe module and the backup state store module.
  • the generic libraries embodiments also eliminate the need for each different application client that is run on each client computing device to separately implement the client state store module, the failure detector module, and the client resource probe module.
  • the decision engine module 126 can be implemented as a generic library that is run on the current data center 104 and is shared by all of the different cloud applications (e.g., 114) that are hosted on the current data center.
  • This particular generic library is hereafter simply referred to as a replication option selection library. All of the applications that are hosted on the current data center can input their performance constraints and resource cost constraints, and their prioritizations thereof, to the replication option selection library. For each of the applications that is hosted on the current data center, the replication option selection library can then use the application's constraints, and it's priorities thereof, to
  • the decision engine module independently selects the optimal session state data replication option for each of the interactive sessions (e.g., 112) that either is currently taking place, or might take place in the future, with the application.
  • the decision engine module as a generic library is particularly well-suited to cloud computing scenarios, virtualized server scenarios, and other scenarios where a plurality of applications are being hosted on the same data center. In these scenarios the current data center can serve a large number of interactive sessions with a large number of applications, and the replication option selection library can be shared amongst all of the sessions and applications.
  • the current state store and current server resource probe modules 128 and 130 can be implemented as another generic library that is run on the current data center 104 and is shared by the various cloud applications (e.g., 114) that are hosted on the current data center.
  • the backup state store and backup server resource probe modules 136 and 134 can be implemented as yet another generic library that is run on each of the backup data centers 106/118 and is shared by the various cloud
  • the client state store, failure detector and client resource probe modules 120/122/124 can be implemented as yet another generic library that each application client 110 can include in its implementation.
  • this library can be implemented in the form of a JavaScript program (which is advantageous since JavaScript programs are supported by most web browsers).
  • this library can be implemented in various platform-specific forms (e.g., on client computing devices that run a Windows® (a registered trademark of Microsoft Corporation) operating system, this library might be implemented in the form of a .NET DLL (dynamic link library). It will be appreciated that both of these generic library implementations are advantageous in that they produce a generic library that is portable across a wide variety of client platforms.
  • this section provides a more detailed description of the different session state data replication options that can be used by the disaster recovery technique embodiments described herein.
  • this section provides a more detailed description of different options for replicating the session state data for the interactive session 112 that is taking place over the network 108 between the application 114 and the application client 110.
  • each of the different session state data replication options has different performance and resource cost trade-offs.
  • TCn is a cost parameter denoting a latency cost to the client computing device that is associated with a given replication option n.
  • BCn is another cost parameter denoting a network bandwidth cost to the client computing device that is associated with the replication option n.
  • BATCn is yet another cost parameter denoting a battery energy drain cost to the client computing device that is associated with the replication option n.
  • BDn is yet another cost parameter denoting a network bandwidth cost to the current data center that is associated with the replication option n.
  • cloud service providers often charge a given cloud application that they are hosting when data that is associated with the application leaves the data center that is hosting the application (such as when the current data center transmits session state data to a client computing device, or to a backup data center).
  • SDn is yet another cost parameter denoting a data storage cost to the current data center that is associated with the replication option n.
  • CDn is yet another cost parameter denoting a processing cost to the current data center that is associated with the replication option n.
  • FIG. 2 summarizes the different session state data replication options that can be used by the disaster recovery technique embodiments described herein and the
  • cost parameters performance and resource cost parameters (hereafter simply referred to as cost parameters) that are associated with each of the replication options.
  • cost parameters performance and resource cost parameters
  • the first of the session state data replication options exemplified in FIG. 2 involves storing the most recent session state data for the interactive session on the client computing device.
  • This particular option is hereafter simply referred to as replication option 1.
  • replication option 1 the most recent session state data for the interactive session is added to each application response message that the application transmits over the network to the application client.
  • the application client Upon the application client receiving a given application response message, the application client will extract the most recent session state data from the application response message, and the client state store module will store this extracted data on the client computing device as described heretofore.
  • the most recent session state data for the interactive session is also added to each of the client request messages that the application client transmits over the network to the application.
  • the application will add the entire set of most recent session state data to each application response message that the application transmits over the network to the application client. As such each application response message that the application client receives from the application will include the entire set of most recent session state data.
  • the application client will also add the entire set of most recent session state data to each client request message that the application client transmits over the network to the application. As such, each client request message that the application receives from the application client will include the entire set of most recent session state data.
  • the copy of the application that receives the client request message will have the entire set of most recent session state data since it is included in the client request message.
  • the application will add the entire set of most recent session state data to the first application response message that the application it transmits to the application client, and to every Nth application response message thereafter that the application transmits to the application client (N being a first prescribed number greater than one).
  • the application client will add the entire set of most recent session state data to the first client request message that the application client transmits to the application, and to every Mth client request message thereafter that the application client transmits to the application (M being a second prescribed number greater than one).
  • M being a second prescribed number greater than one.
  • N can be equal to M, or N can be greater than M, or N can be less than M.
  • the entire set of session state data will be added to the first client request message that the application client transmits to the application and every third client request message thereafter.
  • replication option 1 rather than the entire set of most recent session state data being added to each application response message that the application transmits to the application client, the application will add the entire set of most recent session state data to the first application response message that the application transmits to the application client, and the application will add just the delta in (e.g., just the modified portions of) the session state data to each application response message thereafter that the application transmits to the application client (e.g., the application will add to each application response message after the first response message just the changes in the session state data since it was last transmitted to the application client).
  • the application will add the entire set of most recent session state data to the first application response message that the application transmits to the application client, and the application will add just the delta in (e.g., just the modified portions of) the session state data to each application response message thereafter that the application transmits to the application client (e.g., the application will add to each application response message after the first response message just the changes in the session state data since it was last transmitted to the application client).
  • the application client will add the entire set of most recent session state data to the first client request message that the application client transmits to the application, and the application client will add just the delta in the session state data to each client request message thereafter that the application client transmits to the application (e.g., the application client will add to each client request message after the first client request message just the changes in the session state data since it was last transmitted to the application).
  • replication option 1 rather than the entire set of most recent session state data being added to each application response message that the application transmits to the application client, the application will add the entire set of most recent session state data to the first application response message that the application transmits to the application client, and the application will add just the delta in the session state data to every Nth application response message thereafter that the application transmits to the application client.
  • the application client will add the entire set of most recent session state data to the first client request message that the application client transmits to the application, and the application client will add just the delta in the session state data to every Mth client request message thereafter that the application client transmits to the application.
  • the entire set of session state data will be added to the first client request message that the application client transmits to the application, and just the delta in the session state data will be added to every third client request message thereafter that the application client transmits to the application.
  • the most recent session state data can be added to the client request messages that the application client transmits over the network to the application, and the application response messages that the application transmits over the network to the application client, in various ways.
  • the most recent session state data can be added in the form of a cookie.
  • the most recent session state data can be added into the body of each client request message and each application response message.
  • An alternate embodiment of the disaster recovery technique is also possible where, rather than adding the most recent session state data to the client request and application response messages, the application client can transmit hidden messages that include the most recent session state data to the application, and the application can transmit similar hidden messages to the application client.
  • replication option 1 has certain performance and resource costs that in certain circumstances can make it undesirable compared to others of the session state data replication options exemplified in FIG. 2. More particularly, if the session state data is large (e.g., in the aforementioned example where a user is performing a flight search on a travel website, the session state data may include all of the available flights that meet the search criteria entered by the user thus far), then including the sessions state data in each request message that the application client transmits over the network to the application will slow down the client (and thus negatively impact the client's performance) since the network upload bandwidth that is available to most of the client computing devices will generally be significantly smaller than the network download bandwidth that is available thereto. Furthermore, including the session state data in each client request message that the application client transmits to the application and in each application response message that the application transmits to the application client may increase the request-response latency for the client (and thus negatively impact the client's performance).
  • the cost parameters associated with a given session state data replication option quantify the performance and/or resource utilization overheads associated with implementing the replication option compared to the performance and resource utilization when the replication option is not implemented (e.g., when the session state data is not replicated). For instance, it may be that the application already transmits all the session state data to the application client because the client can perform local processing on the session state data to serve some of the user's requests without having to contact the application. In this case the increase in request-response latency for the client would only be due to the client transmitting the session state data to the application.
  • TCI is a cost parameter denoting the latency cost to the client computing device associated with it receiving (e.g., downloading) the session state data from the application, and transmitting (e.g., uploading) the session state data to the application, in replication option 1.
  • TCI denotes an increase in application client delay associated with this session state data reception and transmission.
  • BC1 is another cost parameter denoting the network bandwidth cost to the client computing device associated with it receiving the session state data from the application, and transmitting the session state data to the application, in replication option 1.
  • BC1 denotes the amount the application client has to pay for this session state data reception and transmission. BC1 may be zero when the client computing device is connected to the network via a wired connection having unlimited data.
  • BC1 may be very high when the client computing device is connected to the network via a conventional cellular data connection operating in roaming mode.
  • BATC1 is yet another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client receiving the session state data from the application, and transmitting the session state data to the application, in replication option 1.
  • B ATC 1 denotes how much battery energy in the client computing device is used for this session state data reception and transmission. While BATC1 may not be very relevant when the client computing device is operating from an AC (alternating current) power source, B ATC 1 is quite relevant when the client computing device is operating from battery power (e.g., when the client computing device is either a smartphone, or a tablet computer, or the like).
  • the second of the session state data replication options exemplified in FIG. 2 involves synchronously storing the most recent session state data for the interactive session in one or more of the backup data centers.
  • This particular option is hereafter simply referred to as replication option 2.
  • the current data center uploads (e.g., transmits) the most recent session state data over the network to one or more of the backup data centers whenever the session state changes, thus ensuring that the backup data centers are storing the most recent session state data.
  • the backup state store module will store this most recent session state data in the one or more backup data centers as described heretofore, and will then transmit a confirmation message to the current data center indicating that the most recent session state data has been successfully stored in the backup data centers.
  • the current data center will upload the most recent session state data over the network to one or more of the backup data centers.
  • the application will then wait to respond to the client request message until the current data center receives the confirmation message from at least one of the backup data centers.
  • the current data center will upload the entire set of most recent session state data to one or more of the backup data centers whenever the session state changes.
  • the current data center will upload the entire set of most recent session state data to one or more of the backup data centers the first time the session state changes, and the current data center will upload just the delta in the session state data to the backup data centers thereafter (e.g., the current data center will upload just the changes in the session state data since it was last uploaded to the backup data centers).
  • Replication option 2 has the following performance and resource cost trade-offs. Since replication option 2 does not burden the client computing device, it avoids any network bandwidth cost thereon. However, replication option 2 does increase the time that the application client has to wait to receive application response messages from the application, which negatively impacts the client's performance.
  • TC2 is a cost parameter denoting the latency cost to the client computing device associated with the application client having to wait to receive application response messages from the application in replication option 2. In other words, TC2 denotes an increase in application client delay associated with this having to wait.
  • Replication option 2 also increases the cost to the current data center for hosting the application since the current data center has to pay for the network bandwidth it uses to upload the most recent session state data to one or more of the backup data centers.
  • BD2 is another cost parameter denoting the network bandwidth cost to the current data center associated with it uploading the session state data to one or more of the backup data centers in replication option 2.
  • BD2 denotes the amount the current data center has to pay for this session state data upload.
  • Replication option 2 also presents the current data center with the cost of storing the session state data in one or more of the backup data centers.
  • SD2 is yet another cost parameter denoting the data storage cost to the current data center associated with it storing the session state data in one or more of the backup data centers in replication option 2.
  • the third of the session state data replication options exemplified in FIG. 2 involves storing an encoded version of the most recent session state data for the interactive session on the client computing device, where the encoding reduces the data size.
  • This particular option is hereafter simply referred to as replication option 3.
  • an encoded version of the most recent session state data for the interactive session is added to each application response message that the application transmits over the network to the application client, and an encoded version of the most recent session state data for the interactive session is added to each of the client request messages that the application client transmits over the network to the application.
  • each of the application response messages that the application client receives from the application will include an encoded version of the most recent session state data.
  • the application client Upon the application client receiving a given application response message, the application client will extract the encoded version of the most recent session state data from the application response message, and the client state store module will store this extracted data on the client computing device as described heretofore.
  • each client request message that the application receives from the application client will include an encoded version of the most recent session state data. Accordingly, in the event that the current data center goes offline and a given client request message from the application client is diverted to the backup data center, the copy of the application that receives the client request message will have an encoded version of the most recent session state data since it is included in the client request message.
  • an encoded version of the most recent session state data can be added to the client request messages and the application response messages in various ways including, but not limited to, the aforementioned various ways that were described in relation to replication option 1.
  • the encoded version of the session state data that is transmitted over the network and stored on the client computing device in replication option 3 is smaller in size than the non-encoded version of the session state data that is transmitted over the network and stored on the client computing device in replication option 1.
  • the encoded version of the session state data may be substantially smaller than the non-encoded version thereof.
  • the encoded version of the most recent session state data includes just the data that the user most recently entered into the client computing device (hereafter simply referred to as most recent user-entered data).
  • the copy of the application that receives the client request message can process the most recent user-entered data and re-create the remainder of the most recent session state data.
  • the encoded version of the session state data is a compressed version of all of the session state data that is transmitted and stored in replication option 1.
  • This compression and the subsequent decompression of the session state data can be performed using any data compression method.
  • a conventional delta compression method is used for this compression and decompression; this particular method is advantageous since, as described heretofore, the session state can change incrementally as user interactions with the application progress.
  • TC3 is a cost parameter denoting the latency cost to the client computing device associated with it receiving the encoded session state data from the application, and transmitting the encoded session state data to the application, in replication option 3. In other words, TC3 denotes an increase in application client delay associated with this encoded session state data reception and transmission.
  • BC3 is another cost parameter denoting the network bandwidth cost to the client computing device associated with it receiving the encoded session state data from the application, and transmitting the encoded session state data to the application, in replication option 3. In other words, BC3 denotes the amount the application client has to pay for this encoded session state data reception and transmission. BC3 may be zero when the client computing device is connected to the network via a wired connection having unlimited data.
  • BC3 may be very high when the client computing device is connected to the network via a cellular data connection operating in roaming mode.
  • BATC3 is yet another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client receiving the encoded session state data from the application, and transmitting the encoded session state data to the application, in replication option 3.
  • BATC3 denotes how much battery energy in the client computing device is used for this encoded session state data reception and transmission. While BATC3 may not be very relevant when the client computing device is operating from an AC power source, BATC3 is quite relevant when the client computing device is operating from battery power.
  • replication option 3 reduces the size of the session state data that is transmitted over the network and stored on the client computing device, TC3 will be less than TCI, BC3 will be less than BC1, and BATC3 will be less than BATC1. In fact, for certain applications TC3, BC3 and BATC3 may be substantially less than TCI, BC1 and BATC1 respectively.
  • CD3 is yet another cost parameter denoting the processing cost to the current data center associated with the application decoding the encoded session state data in replication option 3.
  • the application in the current data center has to decode the encoded session state data in each client request message it receives from the client (e.g., either process the user-entered data and re-create the remainder of the most recent session state data, or decompress the compressed session state data to produce the most recent session state data) in order to be able to respond to the client request message.
  • This decoding operation will increase the processor overhead in the current data center, and will also slightly increase the request-response latency for the client.
  • replication option 4 also involves storing the most recent session state data for the interactive session on the client computing device.
  • Replication option 4 is similar to replication option 1 in that the most recent session state data for the interactive session is added to each application response message that the application transmits over the network to the application client.
  • the application client Upon the application client receiving a given application response message, the application client will extract the most recent session state data from the application response message, and the client state store module will store this extracted data on the client computing device as described heretofore.
  • the most recent session state data can be added to the application response messages in various ways including, but not limited to, the aforementioned various ways that were described in relation to replication option 1.
  • Replication option 4 differs from replication option 1 in the following ways.
  • failure detection software on the client computing device e.g., the aforementioned failure detector module
  • the application client will upload the most recent session state data for the interactive session to a backup data center that has been selected to take over serving the interactive session.
  • the most recent session state data for the interactive session will then be added to each application response message that the copy of the application transmits over the network to the application client.
  • the application client Upon the application client receiving a given application response message from the copy of the application, the application client will extract the most recent session state data from the application response message, and the client state store module will store this extracted data on the client computing device as described heretofore.
  • TC4 is a cost parameter denoting the latency cost to the client computing device associated with it receiving the session state data from the application, and transmitting the session state data to a backup data center whenever the current data center becomes unable to continue serving the interactive session, in replication option 4.
  • TC4 denotes an increase in application client delay associated with this session state data reception and transmission.
  • BC4 is another cost parameter denoting the network bandwidth cost to the client computing device associated with it receiving the session state data from the application, and transmitting the session state data to a backup data center whenever the current data center becomes unable to continue serving the interactive session, in replication option 4.
  • BC4 denotes the amount the application client has to pay for this session state data reception and transmission.
  • BC4 may be zero when the client computing device is connected to the network via a wired connection having unlimited data. Alternatively, BC4 may be very high when the client computing device is connected to the network via a cellular data connection operating in roaming mode.
  • BATC4 is yet another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client receiving the session state data from the application, and the client transmitting the session state data to a backup data center whenever the current data center becomes unable to continue serving the interactive session, in replication option 4. In other words, BATC4 denotes how much battery energy in the client computing device is used for this session state data reception and
  • BATC4 may not be very relevant when the client computing device is operating from an AC power source, BATC4 is quite relevant when the client computing device is operating from battery power. Since replication option 4 does not involve the application client adding the session state data to each client request message the application client transmits over the network to the application (as is done in replication option 1), TC4 will be less than TCI, BC4 will be less than BC1, and BATC4 will be less than BATC1.
  • the fifth of the session state data replication options exemplified in FIG. 2 involves asynchronously storing the most recent session state data for the interactive session in one or more of the backup data centers.
  • This particular option is hereafter simply referred to as replication option 5.
  • the current data center uploads the most recent session state data (either the entire set thereof, or just the delta therein, as described in relation to replication option 2) over the network to one or more of the backup data centers whenever the session state changes, thus ensuring that the backup data centers are storing the most recent session state data.
  • the backup state store module will store this most recent session state data in the one or more backup data centers as described heretofore, and will then transmit a confirmation message to the current data center indicating that the most recent session state data has been successfully stored in the backup data centers.
  • the current data center will upload the most recent session state data over the network to one or more of the backup data centers.
  • replication option 5 the application will respond to the client request message before the current data center receives this confirmation message from at least one of the backup data centers, but the application client will wait to expose the application response to the user of the client computing device until the confirmation message is received from at least one of the backup data centers. Accordingly, in replication option 5, during the time that the most recent session state data is being uploaded to the backup data centers the client computing device is already downloading the application response, thus hiding some of the network round trip time between the current and backup data centers.
  • Replication option 5 can address this risk (e.g., prevent a failure of the current data center from causing the most recent session state data to be lost) in various ways.
  • the application can wait a prescribed period of time after the most recent session state data for the interactive session is uploaded to one or more backup data centers before the application responds to the client request message, where this period of time is made large enough to ensure that the most recent session state data has left the current data center (e.g., has left the current data center's internal network (not shown) and has been transmitted onto the external network).
  • the current data center can add forward error correction (FEC) codes to the most recent session state data before it is uploaded to the backup data centers.
  • FEC forward error correction
  • Replication option 5 has the following performance and resource cost trade-offs. Since replication option 5 does not burden the client computing device, it avoids any network bandwidth cost thereon. However, replication option 5 does very slightly increase the time that the application client has to wait to receive application response messages from the application, which has a very small negative impact on the client's performance.
  • TC5 is a cost parameter denoting the latency cost to the client computing device associated with the application client having to wait to receive application response messages from the application in replication option 5. In other words, TC5 denotes an increase in application client delay associated with the extra amount of time that the application client has to wait to receive application response messages from the application in replication option 5.
  • BATC5 is another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client having to wait this extra amount of time. Given the foregoing, it will be appreciated that TC5 will generally be less than TC2.
  • Replication option 5 increases the cost to the current data center for hosting the application since the current data center has to pay for the network bandwidth it uses to upload the most recent session state data to one or more of the backup data centers.
  • BD5 is yet another cost parameter denoting the network bandwidth cost to the current data center associated with it uploading the session state data and FEC codes to one or more of the backup data centers in replication option 5.
  • BD5 denotes the amount the current data center has to pay for this session state data and FEC codes upload. Due to the addition of the FEC codes to the data that is transmitted, BD5 will be greater than BD2.
  • Replication option 5 also presents the current data center with the cost of storing the session state data in one or more of the backup data centers.
  • SD5 is yet another cost parameter denoting the data storage cost to the current data center associated with it storing the session state data in one or more of the backup data centers in replication option 5. It will be appreciated that SD5 will be the same as SD2.
  • the sixth of the session state data replication options exemplified in FIG. 2 is a variant of replication option 3 that involves storing an encrypted version of the most recent session state data for the interactive session on the client computing device.
  • This particular option is hereafter simply referred to as replication option 6.
  • Replication option 6 is applicable to a situation where the session state data that is computed by the application is to be stored on the client computing device but is not to be disclosed to the user of the client computing device.
  • One example of such a situation is an online card game application where the session state data that is computed by the application includes the hand of cards that is generated for all of the users that are currently playing the game. Encrypting this session state data prevents the hand of cards for users other than the client computing device's user, and the hand of cards for the computerized player, from being disclosed to the client computing device's user.
  • the session state data can be encrypted using any conventional data encryption method.
  • TC6 is a cost parameter denoting the latency cost to the client computing device associated with it receiving the encrypted session state data from the application in replication option 6. In other words, TC6 denotes an increase in application client delay associated with this encrypted session state data reception.
  • BC6 is another cost parameter denoting the network bandwidth cost to the client computing device associated with it receiving the encrypted session state data from the application in replication option 6. In other words, BC6 denotes the amount the application client has to pay for this encrypted session state data reception.
  • BC6 may be zero when the client computing device is connected to the network via a wired connection having unlimited data. Alternatively, BC6 may be very high when the client computing device is connected to the network via a cellular data connection operating in roaming mode.
  • BATC6 is yet another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client receiving the encrypted session state data from the application in replication option 6. In other words, BATC6 denotes how much battery energy in the client computing device is used for this encrypted session state data reception. While BATC6 may not be very relevant when the client computing device is operating from an AC power source, BATC6 is quite relevant when the client computing device is operating from battery power.
  • CD6 is yet another cost parameter denoting the processing cost to the current data center associated with the application encrypting the encoded session state data in replication option 6. This encryption operation will increase the processor overhead in the current data center, and will also slightly increase the request-response latency for the client.
  • the seventh of the session state data replication options exemplified in FIG. 2, hereafter simply referred to as replication option 7, is a hybrid option that can include any combination of replication options 1-6.
  • a session state data replication option can be independently selected for each of the client computing devices.
  • the decision engine module may decide to store the session state data on all of the client computing devices.
  • the decision engine module may also decide to store the session state data on just some of the client computing devices and in one or more of the backup data centers.
  • the decision engine module may also decide to store the session state data on just one or more of the backup data centers.
  • replication option 2 can be used to store the session state data for the interactive session in a backup data center that is near the current data center
  • replication option 5 can be used to store the session state data in another backup data center that is not near (e.g., is far away from) the current data center.
  • Replication option 3 can also be combined with either replication option 2 or replication option 5 so that an encoded version of the session state data for the interactive session is either synchronously or asynchronously stored in each of the backup data centers. As indicated in FIG. 2, various combinations of the different performance and resource cost parameters that are associated with replication options 1-6 become relevant in replication option 7.
  • the different performance and resource cost parameters that are associated with replication options 1-7 might not be static and their values might not be known upfront when a given cloud application is being developed or deployed.
  • a wireless connection such as a conventional cellular data connection, or WiFi connection, or the like (e.g., when the client computing device is either a smartphone, or a tablet computer, or the like)
  • the client's network bandwidth cost e.g., BC1
  • battery energy drain cost e.g., BATC1
  • the client's network bandwidth cost and battery energy drain cost may be negligible.
  • the current data center's network bandwidth cost e.g., BD2
  • data storage cost e.g., SD2
  • the time the application client has to wait to receive application response messages from the application depends on how far the backup data centers are from the current data center.
  • this client wait time may be high since the application may initially be hosted on just two data centers that are located far apart from one another. As the usage of the application grows, the application may be deployed in additional data centers in different locations, or the application may start to rent data center capacity offered by cloud service providers in different locations, either of which may decrease the distance between the current data center and one or more of the backup data centers, thus decreasing the client wait time (e.g., decreasing TC2).
  • the values of the different performance and resource cost parameters that are associated with replication options 1-7 can be different for different interactive sessions with a given cloud application, and can change over time.
  • the disaster recovery technique embodiments described herein can select an appropriate session state data replication option when the application is first deployed since some of the performance and resource cost parameters become known at this time (e.g., the initial number of data centers hosting the application becomes known, the physical locations and thus the distances between the various pairs of the data centers becomes known, and the network bandwidth and data storage costs for each of the data centers become known).
  • the disaster recovery technique embodiments can also select an appropriate session state data replication option whenever there are configuration changes in the computing and networking infrastructure that supports the application (e.g., whenever the number and/or physical locations of data centers hosting the application changes, or whenever the network bandwidth and data storage costs for one or more of the data centers changes).
  • the disaster recovery technique embodiments described herein can also select an appropriate session state data replication option whenever a new interactive session is established between a given cloud application that is running on a given data center and an application client that is running on a given client computing device that is being used by a given user (in this case, all of the performance and resource cost parameters that are associated with replication options 1-7 can be considered).
  • the disaster recovery technique embodiments can also monitor all of the performance and resource cost parameters and again select an appropriate session state data replication option whenever there are changes in these parameters, where this newly selected option can be either the same as or different than the previously selected option.
  • This section provides a more detailed description of the aforementioned various methods that the decision engine module can use to select an appropriate option for replicating the session state data for the interactive cloud application that is being hosted on the current data center.
  • This selection is made from a set of different session state data replication options that includes, but is not limited to, the aforementioned replication options 1-7.
  • each of these replication options has different performance and resource cost trade-offs.
  • this selection can be based on various network performance conditions and resource availability conditions that can be measured by the client resource probe module, current server resource probe module, and backup server resource probe module.
  • This selection can also be based on various performance constraints and resource cost constraints, and a prioritization thereof, that can be specified by the application.
  • FIG. 3 illustrates one embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application that is being hosted on the current data center.
  • the process starts in block 300 with estimating the latency cost to a client computing device associated with each of the session state data replication options in the set of different session state data replication options.
  • TCI, TC2, TC3, TC4, TC5 and TC6 can be estimated using the various measurement techniques described heretofore.
  • the session state data replication option having the smallest estimated latency cost to the client computing device is then selected as the option to be used for replicating the session state data for an interactive session between the application and the client (block 302). It will thus be appreciated that the process embodiment exemplified in FIG. 3 minimizes the latency cost to the client computing device associated with replicating the session state data, and thus minimizes the delay overhead that a user will experience due to the session state data replication.
  • FIG. 4 illustrates another embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application that is being hosted on the current data center.
  • the process starts in block 400 with inputting one or more performance constraints that are specified by the application.
  • performance constraints include a maximum allowable latency cost to a client computing device associated with replicating the session state data for the interactive session (e.g., a constraint on TCI, TC2, TC3, TC4, TC5 and TC6).
  • the performance constraints can optionally also include a minimum allowable availability for the application.
  • One or more resource cost constraints that are specified by the application are then input (block 402).
  • resource cost constraints can include one or more of client resource cost constraints or data center resource cost constraints.
  • client resource cost constraints include, but are not limited to, one or more of a constraint on the network bandwidth cost to the client computing device associated with replicating the session state data for the interactive session (e.g., a constraint on BC1, BC3, BC4 and BC6), or a constraint on the battery energy drain cost to the client computing device associated with replicating this session state data (e.g., a constraint on BATC1, BATC3, BATC4, BATC5 and BATC6), or a constraint on the processing cost to (e.g., a processor utilization budget for) the client computing device associated with replicating this session state data, or a constraint on the memory cost to (e.g., a memory utilization budget for) the client computing device associated with replicating this session state data.
  • a constraint on the network bandwidth cost to the client computing device associated with replicating the session state data for the interactive session e.g., a constraint on BC1, BC3, BC4
  • Exemplary data center resource cost constraints include, but are not limited to, one or more of a constraint on the network bandwidth cost to the current data center associated with replicating the session state data for the interactive session (e.g., a constraint on BD2 and BD5), or a constraint on the data storage cost to the current data center associated with replicating this session state data (e.g., a constraint on SD2 and SD5), or a constraint on the processing cost to the current data center associated with replicating this session state data.
  • a constraint on the network bandwidth cost to the current data center associated with replicating the session state data for the interactive session e.g., a constraint on BD2 and BD5
  • a constraint on the data storage cost to the current data center associated with replicating this session state data e.g., a constraint on SD2 and SD5
  • processing cost to the current data center associated with replicating this session state data.
  • the latency cost to the client computing device associated with each of the session state data replication options in the set of different session state data replication options is estimated (block 404) as described heretofore.
  • the session state data replication options whose estimated latency cost to the client computing device is less than or equal to the just-described maximum allowable latency cost to the client computing device are then identified (block 406).
  • One of these identified options that meets the resource cost constraints is then selected as the option to be used for replicating the session state data for the interactive session (block 408). It will thus be appreciated that the process embodiment exemplified in FIG. 4 can jointly minimize both the latency cost to the client computing device associated with replicating the session state data, and one or more prescribed resource costs that are associated with replicating the session state data.
  • FIG. 5 illustrates yet another embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application that is being hosted on the current data center.
  • the process starts in block 500 with inputting one or more resource cost constraints that are specified by the application.
  • the latency cost to a client computing device associated with each of the session state data replication options in the set of different session state data replication options is then estimated (block 502).
  • the session state data replication options that meet the resource cost constraints are then identified (block 504).
  • the one of these identified options that has the smallest estimated latency cost to the client computing device is then selected as the option to be used for replicating the session state data for the interactive session (block 506).
  • the process embodiment exemplified in FIG. 5 can also jointly minimize both the latency cost to the client computing device associated with replicating the session state data, and one or more prescribed resource costs that are associated with replicating the session state data.
  • the resource cost constraints can be specified by the application in various ways.
  • the resource cost constraints can be specified for individual interactive sessions with the application.
  • the resource cost constraints can also be specified as a total resource cost budget that is to be allocated across a plurality of interactive sessions with the application.
  • the decision engine module can allocate the total resource cost budget in various ways including, but not limited to, the following.
  • the decision engine module can allocate the total resource cost budget on the fly to individual interactive sessions having a current network latency that is higher than a prescribed threshold (e.g., additional resources can be used to reduce the network latency and thus reduce the user response times).
  • the decision engine module can also allocate the total resource cost budget over a prescribed period of time.
  • specific amounts of the budget can be allocated to specific days of the week.
  • the budget can be allocated in proportion to the application load (e.g., the user demand for the application) that is expected over the period of time.
  • different portions of the budget can be allocated to different segments of the period of time based on the application load that is expected during each of the segments.
  • An alternate embodiment of the disaster recovery technique described herein is also possible where the one or more performance constraints that are specified by the application, and the one or more resource cost constraints that are specified by the application, can be combined into a unified constraint by applying a weight to each of the performance constraints and each of the resource cost constraints.
  • weights that are applied to the different performance and resource cost constraints serve to prioritize these constraints.
  • these weights can be used to specify the relative importance to the application of the latency cost to the client computing device as compared to the network bandwidth cost to the client computing device.
  • the weights can be determined in various ways. By way of example but not limitation, in an exemplary implementation of this alternate embodiment the weights are pre-determined by the application.
  • FIG. 6 illustrates an exemplary embodiment, in simplified form, of the server side of a process for ensuring the availability of a cloud application that is being hosted on a current data center.
  • the process starts in block 600 with selecting an option for replicating session state data for the application.
  • this selection is made from a set of different session state data replication options each of which has different performance and resource cost trade-offs, and the selected option determines how the session state data for the application is to be replicated.
  • the selected option is then implemented (block 602), where the implementation results in the session state data for the application being replicated outside of the current data center, thus ensuring that this data remains available in the event that that current data center goes offline.
  • FIG. 7 illustrates an exemplary embodiment, in simplified form, of the client side of a process for providing disaster recovery for the cloud application that is being hosted on the current data center.
  • the process starts in block 700 with receiving a message from the current data center specifying an option for replicating session state data for the application that has been selected for a given client computing device.
  • the specified option is then implemented (block 702), where the implementation results in the session state data for the application being replicated outside of the current data center.
  • the disaster recovery technique embodiments described heretofore assume that a plurality of data centers (e.g., 104/106/118) exist in the cloud 138, and also assume that this plurality of data centers includes one or more backup data centers 106/118 which have sufficient spare (e.g., idle) capacity (e.g., sufficient spare/idle processing capacity and sufficient spare/idle data storage capacity) to take over the current application load of the current data center 104 if it fails.
  • the disaster recovery technique embodiments described heretofore assume that the backup data centers have sufficient spare capacity to take over serving the interactive sessions 112 that are currently taking place with the cloud applications 114 that are currently being hosted by the current data center in the event that it fails.
  • This section describes an exemplary embodiment of the disaster recovery technique described herein that adds spare capacity to each of the data centers in the cloud, where this spare capacity is sufficient to allow a failure of any one of the data centers to be accommodated by the remaining data centers.
  • the spare capacity that is added to each of the remaining data centers by the disaster recovery technique embodiment described in this section allows the current application load of the failed data center to be distributed across the remaining data centers (e.g., the remaining data centers can collectively take over serving the interactive sessions that were being served by the failed data center).
  • FIG. 8 illustrates an exemplary embodiment, in simplified form, of a process for provisioning and using spare data center capacity in the cloud.
  • the process starts in block 800 with determining an amount of spare server capacity (e.g., the number of additional servers) that has to be added to each of the data centers in the cloud in order for the cloud to be able to take over the current application load of any one of the data centers if it fails.
  • This determination is made by taking into account different factors which include, but are not limited to, one or more of the following: the total number of data centers in the cloud, the current server capacity of (e.g., the current number of servers in) each of the data centers, and the geographic location of each of the data centers.
  • the process embodiment exemplified in FIG. 8 is applicable to situations where either the current server capacity of each of the data centers in the cloud in the cloud is either the same, or the current server capacity of one or more of the data centers is different than the current server capacity of others of the data centers.
  • the process embodiment exemplified in FIG. 8 is also applicable to situations where either all of the data centers are located in the same geographic region, or the data centers are located in two or more different geographic regions, where these different geographic regions may be in either the same time zone or different time zones.
  • spare server capacity can be added to each of the data centers, or different amounts of spare server capacity can be added to different ones of the data centers.
  • the process embodiment exemplified in FIG. 8 is advantageous in that it can minimize the amount of spare capacity that has to be added to each of the data centers in the cloud in order for the cloud to be able to take over the current application load of any one of the data centers if it fails. Accordingly, the process embodiment exemplified in FIG. 8 can minimize the expense associated with provisioning spare data center capacity.
  • the disaster recovery technique has been described by specific reference to embodiments thereof, it is understood that variations and modifications thereof can be made without departing from the true spirit and scope of the disaster recovery technique.
  • an embodiment of the disaster recovery technique described herein is possible where the session state data for a given interactive session between a given cloud application and a given application client includes one or more data items that cannot be stored on just the client computing device that is running the application client.
  • HTTPS Hyper Text Transfer Protocol Secure
  • the cryptographic keys associated with the HTTPS session have to be stored on both the application server and the client computing device, and the application and application client already take care of this.
  • just session state data replication options that store the most recent session state data on one or more backup data centers (e.g., either the aforementioned replication option 2 or replication option 5) will be used since the backup data centers have to have the private (e.g., secret) key to even decrypt any client request messages the backup data centers may receive from the application client in the event that the current data center goes offline.
  • the client state store module will encrypt this extracted data using any conventional data encryption method, and will then store this encrypted data and an associated encryption key on the client computing device.
  • the session resume request message that the application client transmits to the backup data center that has been selected to take over serving the interactive session will include both the encrypted most recent session state data for the interactive session and the associated encryption key that are stored on the client computing device.
  • This particular embodiment is applicable in situations where the session state data includes proprietary information that may not be appropriate to expose to a user. Examples of such proprietary information include customer profiling data in an customer database, a selection of advertisements to be shown or discount codes to be offered during the session, certain types of information generated by an ecommerce application, and the like. Additionally, in an online card game application the session state data may include the hand of cards that is generated for all of the users that are currently playing the game. Storing such proprietary information in encrypted form on the client computing device is advantageous since the information is kept secure.
  • FIG. 9 illustrates a simplified example of a general- purpose computer system on which various embodiments and elements of the disaster recovery technique, as described herein, may be implemented. It is noted that any boxes that are represented by broken or dashed lines in the simplified computing device 900 shown in FIG. 9 represent alternate embodiments of the simplified computing device. As described below, any or all of these alternate embodiments may be used in combination with other alternate embodiments that are described throughout this document.
  • the simplified computing device 900 is typically found in devices having at least some minimum computational capability such as personal computers (PCs), server computers, handheld computing devices, laptop or mobile computers, communications devices such as cell phones and personal digital assistants (PDAs), multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, and audio or video media players.
  • PCs personal computers
  • server computers handheld computing devices
  • laptop or mobile computers such as cell phones and personal digital assistants (PDAs)
  • PDAs personal digital assistants
  • multiprocessor systems microprocessor-based systems
  • set top boxes programmable consumer electronics
  • network PCs network PCs
  • minicomputers minicomputers
  • mainframe computers mainframe computers
  • audio or video media players audio or video media players
  • the device should have a sufficient computational capability and system memory to enable basic computational operations.
  • the computational capability of the simplified computing device 900 shown in FIG. 9 is generally illustrated by one or more processing unit(s) 910, and may also include one or more graphics processing units (GPUs) 915, either or both in communication with system memory 920.
  • the processing unit(s) 910 of the simplified computing device 900 may be specialized microprocessors (such as a digital signal processor (DSP), a very long instruction word (VLIW) processor, a field-programmable gate array (FPGA), or other micro-controller) or can be conventional central processing units (CPUs) having one or more processing cores.
  • DSP digital signal processor
  • VLIW very long instruction word
  • FPGA field-programmable gate array
  • CPUs central processing units having one or more processing cores.
  • the simplified computing device 900 shown in FIG. 9 may also include other components such as a communications interface 930.
  • the simplified computing device 900 may also include one or more conventional computer input devices 940 (e.g., pointing devices, keyboards, audio (e.g., voice) input devices, video input devices, haptic input devices, gesture recognition devices, devices for receiving wired or wireless data transmissions, and the like).
  • the simplified computing device 900 may also include other optional components such as one or more conventional computer output devices 950 (e.g., display device(s) 955, audio output devices, video output devices, devices for transmitting wired or wireless data transmissions, and the like).
  • typical communications interfaces 930, input devices 940, output devices 950, and storage devices 960 for general-purpose computers are well known to those skilled in the art, and will not be described in detail herein.
  • the simplified computing device 900 shown in FIG. 9 may also include a variety of computer-readable media.
  • Computer-readable media can be any available media that can be accessed by the computer 900 via storage devices 960, and can include both volatile and nonvolatile media that is either removable 970 and/or non-removable 980, for storage of information such as computer-readable or computer-executable instructions, data structures, program modules, or other data.
  • Computer-readable media includes computer storage media and communication media.
  • Computer storage media refers to tangible computer-readable or machine-readable media or storage devices such as digital versatile disks (DVDs), compact discs (CDs), floppy disks, tape drives, hard drives, optical drives, solid state memory devices, random access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, magnetic cassettes, magnetic tapes, magnetic disk storage, or other magnetic storage devices.
  • DVDs digital versatile disks
  • CDs compact discs
  • floppy disks tape drives
  • hard drives optical drives
  • solid state memory devices random access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, magnetic cassettes, magnetic tapes, magnetic disk storage, or other magnetic storage devices.
  • RAM random access memory
  • ROM read-only memory
  • EEPROM electrically erasable programmable read-only memory
  • flash memory or other memory technology
  • magnetic cassettes magnetic tapes
  • Retention of information such as computer-readable or computer-executable instructions, data structures, program modules, and the like, can also be accomplished by using any of a variety of the aforementioned communication media (as opposed to computer storage media) to encode one or more modulated data signals or carrier waves, or other transport mechanisms or communications protocols, and can include any wired or wireless information delivery mechanism.
  • modulated data signal or “carrier wave” generally refer to a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
  • communication media can include wired media such as a wired network or direct-wired connection carrying one or more modulated data signals, and wireless media such as acoustic, radio frequency (RF), infrared, laser, and other wireless media for transmitting and/or receiving one or more modulated data signals or carrier waves.
  • wired media such as a wired network or direct-wired connection carrying one or more modulated data signals
  • wireless media such as acoustic, radio frequency (RF), infrared, laser, and other wireless media for transmitting and/or receiving one or more modulated data signals or carrier waves.
  • RF radio frequency
  • the disaster recovery technique embodiments described herein may be further described in the general context of computer-executable instructions, such as program modules, being executed by a computing device.
  • program modules include routines, programs, objects, components, data structures, and the like, that perform particular tasks or implement particular abstract data types.
  • the disaster recovery technique embodiments may also be practiced in distributed computing environments where tasks are performed by one or more remote processing devices, or within a cloud of one or more devices, that are linked through one or more communications networks.
  • program modules may be located in both local and remote computer storage media including media storage devices.
  • the aforementioned instructions may be implemented, in part or in whole, as hardware logic circuits, which may or may not include a processor.

Abstract

Disaster recovery is provided for an application that is being hosted on a current data center, thus ensuring the availability of the application. An option for replicating session state data for the application is selected. This selection is made from a set of different session state data replication options each of which has different performance and resource cost trade-offs. The selected option determines how the session state data for the application is to be replicated. The selected option is implemented, where the implementation results in the session state data for the application being replicated outside of the current data center, thus ensuring that this data remains available in the event that the current data center goes offline.

Description

GEO-DISTRIBUTED DISASTER RECOVERY FOR INTERACTIVE CLOUD
APPLICATIONS
BACKGROUND
[01] The Internet is a global data communications system that serves billions of users worldwide. The Internet provides users access to a vast array of online information resources and services, including those provided by the World Wide Web, intranet-based enterprises, and the like. Thanks to the ubiquity of the various types of personal computing devices that exist today (such as personal computers, laptop/notebook computers, smart phones and tablet computers) and the ubiquity of the Internet, an evergrowing number of users in an increasing number of geographic regions routinely use an ever-growing number of web applications in their everyday lives. For examples, users rely on web applications for emailing, texting, video conferencing, listening to music, watching movies, searching for online information, storing and accessing online information, and online shopping and banking. Web applications are commonly hosted on data centers.
SUMMARY
[02] This Summary is provided to introduce a selection of concepts, in a simplified form, that are further described hereafter in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
[03] Geo-distributed disaster recovery technique embodiments described herein are generally applicable to ensuring the availability of an application that is being hosted on a current data center. In other words, the geo-distributed disaster recovery technique embodiments provide disaster recovery for the application. In one exemplary embodiment an option for replicating session state data for the application is selected, where this selection is made from a set of different session state data replication options each of which has different performance and resource cost trade-offs, and the selected option determines how the session state data for the application is to be replicated. The selected option is then implemented, where the implementation results in the session state data for the application being replicated outside of the current data center, thus ensuring that this data remains available in the event that the current data center goes offline.
[04] In another exemplary embodiment a client computer receives a message from the current data center specifying an option for replicating session state data for the application that has been selected for the client computer, where this selection has been made from a set of different session state data replication options each of which has different performance and resource cost trade-offs, and the specified option determines how the session state data for the application is to be replicated. The specified option is then implemented, where the implementation results in the session state data for the application being replicated outside of the current data center, thus ensuring that this data remains available to the client computer in the event that the current data center goes offline.
[05] Yet another exemplary embodiment involves provisioning and using spare data center capacity in a network of data centers. A determination is made as to an amount of spare server capacity that is to be added to each of the data centers in the network of data centers in order for this network to be able to take over the current application load of any one of the data centers if it fails. This determination is made by taking into account different factors that include one or more of: the total number of data centers, or the current server capacity of each of the data centers, or the geographic location of each of the data centers. It is then directed that the determined amount of spare server capacity be added to each of the data centers. Then, whenever any one of the data centers fails, the current application load of the failed data center is distributed across the spare server capacity that was added to each of the remaining data centers.
BRIEF DESCRIPTION OF THE DRAWINGS
[06] The specific features, aspects, and advantages of the geo-distributed disaster recovery technique embodiments described herein will become better understood with regard to the following description, appended claims, and accompanying drawings where:
[07] FIG. 1 is a diagram illustrating an exemplary embodiment, in simplified form, of an architectural framework for implementing the geo-distributed disaster recovery technique embodiments described herein.
[08] FIG. 2 is a table summarizing different session state data replication options that can be used by the geo-distributed disaster recovery technique embodiments described herein and the performance and resource cost parameters that are associated with each of the replication options.
[09] FIG. 3 is a flow diagram illustrating one embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for a cloud application that is being hosted on a data center. [010] FIG. 4 is a flow diagram illustrating another embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application.
[Oil] FIG. 5 is a flow diagram illustrating yet another embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application.
[012] FIG. 6 is a flow diagram illustrating an exemplary embodiment, in simplified form, of the server side of a process for ensuring the availability of the cloud application.
[013] FIG. 7 is a flow diagram illustrating an exemplary embodiment, in simplified form, of the client side of a process for providing disaster recovery for the cloud application.
[014] FIG. 8 is a flow diagram illustrating an exemplary embodiment, in simplified form, of a process for provisioning and using spare data center capacity in the cloud.
[015] FIG. 9 is a diagram illustrating a simplified example of a general-purpose computer system on which various embodiments and elements of the geo-distributed disaster recovery technique, as described herein, may be implemented.
DETAILED DESCRIPTION
[016] In the following description of geo-distributed disaster recovery technique embodiments (hereafter simply referred to as disaster recovery technique embodiments) reference is made to the accompanying drawings which form a part hereof, and in which are shown, by way of illustration, specific embodiments in which the disaster recovery technique can be practiced. It is understood that other embodiments can be utilized and structural changes can be made without departing from the scope of the disaster recovery technique embodiments.
[017] It is also noted that for the sake of clarity specific terminology will be resorted to in describing the disaster recovery technique embodiments described herein and it is not intended for these embodiments to be limited to the specific terms so chosen.
Furthermore, it is to be understood that each specific term includes all its technical equivalents that operate in a broadly similar manner to achieve a similar purpose.
Reference herein to "one embodiment", or "another embodiment", or an "exemplary embodiment", or an "alternate embodiment", or "one implementation", or "another implementation", or an "exemplary implementation", or an "alternate implementation" means that a particular feature, a particular structure, or particular characteristics described in connection with the embodiment or implementation can be included in at least one embodiment of the disaster recovery technique. The appearances of the phrases "in one embodiment", "in another embodiment", "in an exemplary embodiment", "in an alternate embodiment", "in one implementation", "in another implementation", "in an exemplary implementation", and "in an alternate implementation" in various places in the
specification are not necessarily all referring to the same embodiment or implementation, nor are separate or alternative embodiments/implementations mutually exclusive of other embodiments/implementations. Yet furthermore, the order of process flow representing one or more embodiments or implementations of the disaster recovery technique does not inherently indicate any particular order not imply any limitations of the disaster recovery technique.
[018] The term "cloud application" is used herein to refer to a web application that operates in the cloud and can be hosted on (e.g., deployed at) a plurality of data centers that can be located in different geographic regions (e.g., different regions of the world) and can be concurrently used by a plurality of remote end users (hereafter simply referred to as users). In other words, a cloud application is a web application that is geographically distributed. The term "client computing device" is used herein to refer to a computing device that a user uses to establish an interactive session with a given a given interactive cloud application.
1.0 Geo-Distributed Disaster Recovery for Interactive Cloud Applications
[019] As is appreciated in the art of cloud computing, there is an emerging class of interactive cloud applications that are being rapidly developed and hosted on popular cloud platforms. Generally speaking, the disaster recovery technique embodiments described herein provide geo-distributed disaster recovery for such applications, while at the same time minimizing the performance costs (e.g., minimizing client latencies) and minimizing the resource costs (e.g., network bandwidth costs and data storage costs, among others) associated with providing this disaster recovery. In other words, the disaster recovery technique embodiments generally involve ensuring (e.g., maximizing) the availability of interactive cloud applications so that the applications and their most recent (e.g., freshest) state data remain available to users even if one or more of the data centers that host the applications go offline for whatever reason.
[020] Generally speaking and as is appreciated in the art of web applications, there are two types of data that are associated with the overall state of a given interactive cloud application, namely session state data and persistent state data. Session state data is associated with an interactive session that one or more users establish with the application. It will be appreciated that the session state can change incrementally each time a user interacts with the application. Persistent state data has two properties that the session state data does not have namely durability and concurrency. The durability property of persistent state data refers to the fact that the persistent state data persists after all of the users have left the interactive session, and after the interactive session ends. In contrast, the session state data persists just for the life of the interactive session, and can be discarded once all of the users have left the interactive session. The concurrency property of persistent state data refers to the fact that a plurality of users and a plurality of interactive sessions can use the persistent state data at the same time. Many of the user's interactions with the application affect the session state data but do not affect the persistent state data.
[021] As will be appreciated from the more detailed description that follows, all of the disaster recovery technique embodiments described herein focus specifically on replicating the session state data for an interactive session that takes place over a network between a given interactive cloud application and one or more client computing devices. However, some of the disaster recovery technique embodiments can also be used to replicate persistent state data for the application. The disaster recovery technique embodiments ensure that the most recent session state data will remain available to the client computing devices in the event that the data center that is currently hosting the application and serving the session goes offline and thus becomes unable to continue serving the session. As is appreciated in the art of cloud computing, various techniques exist for protecting the persistent state data in the event of a data center failure, one example of which is the conventional Paxos family of protocols. The fact that the disaster recovery technique embodiments treat the session state data separately from the persistent state data does not require significant modifications to existing web applications because such applications already preserve the session state explicitly between successive client requests by using a session state application programming interface (API) that is provided by the development platform for the web applications. The disaster recovery technique embodiments maintain consistency between the session state data and persistent state data since the embodiments store and will recover the most recent session state data whenever the data center that is currently serving the session goes offline for whatever reason, and thus becomes unable to continue serving the session.
[022] As will be described in more detail hereafter, there are various performance and resource cost (e.g., resource utilization) trade-offs that are associated with ensuring the availability of interactive cloud applications. The disaster recovery technique embodiments allow developers of a given interactive cloud application to ensure the availability of the application to users by replicating the session state data for the application. As such, the session state data for interactive sessions with the application is protected against data center failures. As will be appreciated from the more detailed description that follows, the disaster recovery technique embodiments replicate the entire session state, including both user-entered data and server-computed data.
[023] As will also be described in more detail hereafter, the disaster recovery technique embodiments provide a set of different options for replicating the session state data for the application, where each of the different session state data replication options has different performance and resource cost trade-offs. The disaster recovery technique embodiments also provide various methods for automatically selecting an appropriate one of the session state data replication options that increases (e.g., maximizes) the performance of a given interactive session with the application (and hence the performance that the users of the application experience) and/or decreases (e.g., minimizes) the resource costs (e.g., decreases the resources that are utilized) for the interactive session. This automatic selection is based on various performance constraints and resource cost constraints, and a prioritization thereof, that can be specified by the application. The disaster recovery technique embodiments can automatically switch from a previously selected session state data replication option to a newly selected session state data replication option in the middle of the interactive session based on changes in these constraints, or changes in their prioritization, or changes in prescribed run-time conditions in the computing and networking infrastructure that supports the application.
[024] In other words, the disaster recovery technique embodiments automatically tune the performance and resource cost trade-offs that are associated with ensuring the availability of interactive cloud applications, and dynamically adapt this tuning (e.g., adapt the tuning on-the-fly) whenever the just-described prescribed run-time conditions change. Generally speaking and as will also be described in more detail hereafter, these run-time conditions can include, but are not limited to, various performance and resource costs such as network bandwidth costs, network latency experienced by client computing devices, data storage costs, processing costs, and battery energy drains costs, among others. The run-time conditions can also include, but are not limited to, various resource availabilities such as the current availability (e.g., percent utilization) of computing device processors, memory, and batteries, among others. Each of these run-time conditions can be different for different client and server computing devices in the infrastructure, and for different portions of the network in the infrastructure. Each of the run-time conditions can also change over time.
[025] The disaster recovery technique embodiments described herein are advantageous for various reasons including, but not limited to, the following. As will be appreciated from the more detailed description that follows, the disaster recovery technique embodiments make interactive cloud applications resilient to data center failures by replicating the session state data for the applications. The disaster recovery technique embodiments minimize the computing and network transmission costs associated with this session state data replication so that the performance of interactive sessions with the applications is maximized, the user response times for the applications are minimized, and the resource costs for the data centers that host the applications are minimized. The disaster recovery technique embodiments also allow regional legal requirements to be met (e.g., current regulations in some countries dictate that some of the data that is served by a data center be located within national boundaries). The disaster recovery technique embodiments also integrate easily with existing cloud applications, in some cases without source code modifications. The disaster recovery technique embodiments automatically adapt to network bandwidth variations at the client computing devices (e.g., network bandwidth variations caused by either cellular or WiFi network signal strength
fluctuations, or an Internet service provider's cable loop utilization, or DSL (digital subscriber line) network wire length variations, or edge network congestion (also known as last mile congestion), or the fact that mobile proxies may compress data sent to the client computing devices, or the like). The disaster recovery technique embodiments also automatically adapt to variable data center placements and transparently benefit cloud applications as new data centers are added to the cloud.
[026] As described heretofore, an ever-growing number of users in an increasing number of geographic regions routinely use an ever-growing number of web applications in their everyday lives. This situation presents the data centers that host web applications with scalability challenges due to real world limitations on the capacity (e.g., storage size, computing power, network bandwidth, and power consumption) of the data centers. The disaster recovery technique embodiments described herein are further advantageous in that they provide usage scalability for interactive cloud applications (e.g., the disaster recovery technique embodiments ensure the availability of a given interactive cloud application even if a large number of users is using the application at the same time). The disaster recovery technique embodiments described herein also allow interactive sessions that take place over a network between interactive cloud applications and client computing devices to be served by nearby data centers. In other words, the disaster recovery technique embodiments ensure that the data center that is hosting a given interactive cloud application and currently serving a given interactive session between the application and a given client computing device will be close to the client computing device regardless of which geographic region the client computing device is in. This close proximity of the data center and client computing device reduces client latency and user response times.
[027] It will be appreciated that a user of a given interactive cloud application is particularly sensitive to the performance of the application during the interactive phases of a session with the application. By way of example but not limitation, a user of an ecommerce website may want fast response time when browsing for products on the website, but may accept slower response times during the final stage of making a product purchase involving credit card verification. Once a user starts using the application (e.g., once the user enters a search request for a desired product name on an ecommerce website, or enters desired departure and destination cities and desired dates for a flight search on a travel website, or invites a friend to an online chat session, or the like), the disaster recovery technique embodiments described herein assume that an interactive session with the application has begun. Data that is relevant to the interactive session can be routinely generated by both the user and the application as the session progresses. By way of example but not limitation, once the flight search on the travel website has been completed, the website may have generated data about which available flights operate between the desired departure and destination cities on the desired dates. The user may have also entered filters for the flight search (e.g., the user may have requested to see just available flights that depart after 5:00 PM and have at most one stop, among other types of filters). At this point in the interactive session the session state data would include both the data that was entered by the user (e.g., the desired departure and destination cities, the desired dates, and the filters) and the data that was generated by the website in response to this user-entered data (e.g., which available flights operate between the desired departure and destination cities on the desired dates and comply with the filters).
[028] The disaster recovery technique embodiments described herein ensure that the most recent session state data remains available to a given interactive cloud application even if the data center that is hosting the application and currently serving an interactive session over a network between the application and one or more client computing devices goes offline for any reason and another data center that is also hosting the application takes over serving the session. In other words, in the just-described example where the user is performing the flight search on the travel website, if the data center that is hosting the application and currently serving the interactive session with the user goes offline, the disaster recovery technique embodiments ensure that the other data center that takes over serving the session has access to the most recent session state data. Accordingly, the user will not have to re-enter any of data that they previously entered, and the other data center will not have to re-compute any of the data that was previously computed in response to the user-entered data. If the user then enters another filter that requests to see just the available flights that operate through a particular connecting airport, this other filter will be routed to the other data center, which will use the most recent session state data to generate new data for the user indicating which of the previously selected and filtered flights comply with this other filter.
1.1 Architectural Framework
[029] FIG. 1 illustrates an exemplary embodiment, in simplified form, of an architectural framework for implementing the disaster recovery technique embodiments described herein. As exemplified in FIG. 1, the architectural framework 100 includes one or more client computing devices 102/116, a network 108, and the cloud 138. As is appreciated in the art of cloud computing, the cloud 138 generally includes a plurality of data centers (e.g., 104/106/118) that can be located in different geographic regions. The network 108 allows the client computing devices 102/116 to communicate with and exchange information with the data centers 104/106/118, and also allows the data centers to communicate with and exchange information with each other. It will be appreciated that the network 108 can employ a wide variety of local and wide area networking
technologies. By way of example but not limitation, in an exemplary embodiment of the disaster recovery technique the network 108 is the Internet. Each of the client computing devices 102/116 can be any type of computing device. By way of example but not limitation, each of the client computing devices 102/116 can be either a mobile computing device (such as a smartphone, or a tablet computer, or a personal digital assistant, or the like) or a non-mobile computing device (such as a desktop personal computer (PC), or a laptop computer, or the like).
[030] As described heretofore and referring again to FIG. 1, users (not shown) can use the client computing devices 102/116 to establish an interactive session 112 with a desired interactive cloud application 114/132 (hereafter sometimes simply referred to as a cloud application, or sometimes even more simply referred to as an application) that can be hosted on the data centers 104/106/118. More particularly, when a given user who is using a given client computing device 102 wants to use a desired cloud application 114/132, various things happen including, but not limited to, the following. An
application client 110 that is associated with the application 114/132 is run on the client computing device 102. One of the data centers 104 that is hosting the application 114 is selected to serve an interactive session 112 between the application and the application client 110. As exemplified in FIG. 1, the data center that is currently selected to serve the interactive session 112 is herein referred to as the current data center 104.
[031] Referring again to FIG. 1, one or more other data centers 106/118 that host a copy of the application 132 are selected to serve the interactive session 112 in the event that the current data center 104 becomes unable to do so for whatever reason. It will be appreciated that the current data center 104 can become unable to continue serving the interactive session 112 for various reasons including, but not limited to, the following. The current data center 104 can go offline due to either a failure within the current data center or a failure in the network 108, or the current data can become overloaded. As exemplified in FIG. 1, the other data centers that are selected to serve the interactive session 112 in the event that the current data center 104 becomes unable to do so are herein referred to as backup data centers 106/118. The disaster recovery technique embodiments described herein assume that the copy of the application 132 that is hosted on the backup data centers 106/118 can provide the same functionality as the application 114 that is hosted on the current data center 104.
[032] As is appreciated in the art of cloud computing and referring again to FIG. 1, each of the data centers 104/106/118 in the architectural framework 100 can include a plurality of server computing devices (not shown) which communicate over an internal network (not shown) in the data center. Generally speaking, large data centers include a large number of server computing devices, and small data centers include a small number of server computing devices. The disaster recovery technique embodiments described herein can be used in various framework 100 configurations ranging from a configuration that includes a small number of large data centers to a configuration that includes a large number of small data centers. In an exemplary embodiment of the disaster recovery technique the application 114 is run on one of the server computing devices in the current data center 104. In other words, the interactive session 112 that is taking place over the network 108 between the application 114 and the application client 110 is served by one of the server computing devices in the current data center 104. The server computing device in the current data center 104 that runs the application 114 and serves the interactive session 112 is hereafter simply referred to as an application server.
[033] As is also appreciated in the art of cloud computing and referring again to FIG. 1, the selection of which one of the data centers 104/106/118 that hosts the application 114/132 will be the current data center 104, and which one or more of these data centers will be backup data centers 106/118, can be made dynamically (e.g., on-the-fly) using various conventional methods. By way of example but not limitation, this selection can be made by either a network of edge nodes (not shown) that is deployed for the application 114/132, or a content delivery network (also known as a content distribution network or CDN for short) (not shown), or by the application client 110 itself (as is the case when using a round robin Domain Name System). Regardless of the specific method that is used to select the current data center 104 and the backup data center(s) 106/118, the disaster recovery technique embodiments described herein assume that these selections have already been made. Once the current data center 104 and the backup data center(s) 106/118 have been selected, an interactive session 112 can be established over the network 108 between the application client 110 and the application 114 that is running in the current data center. The disaster recovery technique embodiments assume that this preselection of the current data center 104 and the backup data center(s) 106/118 is known for each interactive session 112.
[034] Referring again to FIG. 1, if the current data center 104 goes offline at a time when there are a plurality of backup data centers 106/118 in operation, the determination of which of the backup data centers will be selected to become the new current data center for each existing and new interactive session 112 can be made in various ways. By way of example but not limitation, in one embodiment of the disaster recovery technique described herein this selection can be made based on static provisioning decisions (e.g., if two backup data centers 106 and 118 are currently in operation, one of these backup data centers 106 can take over serving one half of the existing and new interactive sessions 112, and the other of these backup data centers 118 can take over serving the other half of the existing and new interactive sessions). In another embodiment of the disaster recovery technique this selection can be made based on the current load conditions in each of the backup data centers 106/118 when the interactive session 112 is first established.
[035] Referring again to FIG. 1, it is noted that during the interactive session 112, application-related messages are exchanged over the network 108 just between the application client 110 that is running on the client computing device 102 and the application 114 that is running in the current data center 104. In other words, no application-related messages are exchanged between the current data center 104 and the backup data centers 106/118. As will be appreciated from the more detailed description that follows and as exemplified in FIG. 1, in addition to the application-related messages, the various additional program modules that are described in more detail hereafter can exchange various other types of information over the network 108 between the client computing devices 102/116 and the current data center 104, and between the current data center and the backup data centers 106/118. These particular information exchanges are hereafter simply referred to as disaster recovery information exchanges.
[036] It will be appreciated that the just-described disaster recovery information exchanges can be implemented in various ways. By way of example but not limitation, in instances where a given disaster recovery information exchange is to take place between a given client computing device (e.g., 102) and the current data center 104, the disaster recovery information exchange can be piggybacked on (e.g., included within) a given application-related message that is to be exchanged over the network 108 between the client computing device and the current data center. In instances where a given disaster recovery information exchange is to take place between the current data center 104 and a given backup data center (e.g., 106), the disaster recovery information exchange can be implemented by the additional program modules transmitting their own messages over the network 108 which are independent of the application-related messages. It is also possible for disaster recovery information exchanges between a given client computing device (e.g., 102) and the current data center 104 to be implemented by the additional program modules transmitting their own messages over the network 108 which are independent of the application-related messages.
[037] Referring again to FIG. 1, in addition to the application client 110, various additional program modules can be run on each of the client computing devices 102/116. As exemplified in FIG. 1, these additional program modules include, but are not limited to, a client state store module 120, a failure detector module 122, and a client resource probe module 124. Exemplary functions of each of these additional program modules 120/122/124 are described in more detail hereafter. As will be appreciated from the more detailed description of the different session state data replication options that also follows, some of these options necessitate the use of all of the additional program modules 120/122/124, and others of these options necessitate the use of just a subset of the additional program modules. Hence, some of the client computing devices 102/116 may be running all of the additional program modules 120/122/124 while other client computing devices may be running just a subset of the additional program modules. The particular session state data replication option that is currently selected and being implemented for a given client computing device (e.g., 102) determines which of the additional program modules 120/122/124 are running on the client computing device at a given point in time.
[038] Referring again to FIG. 1, the client state store module 120 stores session state data for the interactive session 112 on the client computing device 102 for the session state data replication options that involve storing the session state data on the client computing device. In other words, the client state store module inputs a data object (not shown) that includes the most recent session state data and stores it on the client computing device 102. The client state store module can also retrieve and output this stored data object whenever the application client 110 requests it. The client state store module thus ensures that the most recent session state data remains available to the application client as long as the client computing device is available (e.g., remains functional).
[039] Referring again to FIG. 1, the client state store module 120 can be implemented in various ways. By way of example but not limitation, in one embodiment of the disaster recovery technique described herein the client state store module can use the cookie mechanism that is offered by conventional web browsers to store data on the client computing device 102. In another embodiment of the disaster recovery technique the client state store module can use a JavaScript program that stores the data object on the client computing device in either volatile memory or various types of non- volatile memory (e.g., persistent storage) that may be available on the client computing device. Examples of such non- volatile memory include, but are not limited to magnetic disk storage and flash memory. It will be appreciated that storing the most recent session state data in volatile memory on the client computing device is not problematic since, as described heretofore, the session state data persists just for the life of the interactive session, and can be discarded once the application client 110 has left the interactive session. In yet another embodiment of the disaster recovery technique the client state store module can use the conventional web storage APIs that are supported by today's web browser applications to write the most recent session state data to secondary storage such as a blob storage element in the cloud 138. This particular embodiment is advantageous in that it minimizes memory overhead on the client computing device. [040] Referring again to FIG. 1 and as will be described in more detail hereafter, the failure detector module 122 operates when the session state data for the interactive session 112 is being replicated using a fourth one of the different session state data replication options. Generally speaking, the failure detector module operates in two phases, namely a failure detection phase and a subsequent recovery phase. As will be appreciated from the more detailed description that follows, the features of the failure detector module are advantageous in that they minimize the time it takes to recover from the current data center 104 going offline. The disaster recovery technique embodiments described herein assume a conventional fail-stop failure model for the various data centers 104/106/118 in the cloud 138. As is appreciated in the art of distributed computing, due to the well- known Impossibility of Distributed Consensus with One Faulty Process (also known as the "FLP result" and the "FLP impossibility proof), a distributed failure detector has to choose between completeness (e.g., every failure is eventually detected) and accuracy (e.g., false positives). The disaster recovery technique embodiments described herein provide a failure detection solution that is fast and ensures completeness, but does not necessarily ensure accuracy. It is noted that alternate embodiments of the disaster recovery technique are possible that are more complex but can ensure both completeness and accuracy.
[041] During the failure detection phase the failure detector module 122 detects when the current data center 104 that is serving the interactive session 112 goes offline and thus becomes unable to continue serving the session. The failure detector module can be implemented in various ways. By way of example but not limitation, in one embodiment of the disaster recovery technique described herein the failure detector module can use a timer to measure the time that has elapsed since the application client 110 last submitted a client request message over the network 108 to the application 114. If the application client does not receive an application response to the client request message within a prescribed period of time, then the failure detector module can assume that the current data center has failed and gone offline (although it will be appreciated that the actual failure that resulted in the current data center's inability to respond could be somewhere in the network or elsewhere). In another embodiment of the disaster recovery technique the failure detector module can periodically transmit a test message over the network to the current data center and receive a test response message therefrom that indicates the current availability thereof (e.g., this test response message may indicate that the current data center is either fully available, or is currently getting overloaded and hence may not be available in the near future, or is currently operating from battery backup power and hence may fail in the near future, among other things). In yet another embodiment of the disaster recovery technique, whenever the failure detector module detects a failure of the current data center that results in it being currently unable to serve the interactive session, the failure detector module can periodically test the availability of the network at the client computing device 102 using a mechanism that is independent of the current data center, and the failure detector module can use this network availability information to enhance its knowledge about why the current data center is currently unable to serve the interactive session.
[042] Referring again to FIG. 1, once the failure detection phase is complete, the recovery phase is started by the application client 110 transmitting a session resume request message over the network 108 to a backup data center (e.g., 106) that has been selected to take over serving the interactive session 112, where this message includes the most recent session state data for the interactive session that is stored on the client computing device 102. When the selected backup data center receives this session resume request message from the application client, the selected backup data center will extract the most recent session state data from the message and provide it to the copy of the application 132 so that the interactive session 112 can be resumed.
[043] It will be appreciated that after the current data center goes offline, the application is paused until the most recent session state data has been recovered and the interactive session has been resumed at the selected backup data center. This pausing of the application can make the user think that the application has frozen. In order to minimize the user's frustration and prevent them from thinking that the application has frozen, after the failure detector module detects that the current data center has gone offline the failure detector module can optionally inform the application client that the current data center has gone offline, and the application client can optionally display a message to the user telling them about this and that a wait is expected until the interactive session is resumed. Once the recovery phase has been successfully completed and the interactive session has been resumed at a backup data center, the application client can optionally display another message to the user telling them that the application is once again available for use.
[044] Generally speaking and referring again to FIG. 1, the client resource probe module 124 operates cooperatively with a current server resource probe module 130 that is running on the current data center 104 to periodically measure relevant network performance conditions and resource availability conditions at the client computing device 102. A decision engine module 126 that is also running on the current data center 104 can use these measured conditions, along with other measured conditions described herein, to select an appropriate session state data replication option. More particularly and by way of example but not limitation, in one embodiment of the disaster recovery technique described herein the client resource probe module 124 can receive server probe messages that are periodically sent to the client resource probe module by the current server resource probe module 130, and can immediately respond to each of these messages. The current server resource probe module 130 can measure the current network latency associated with communicating with the client computing device 102 (e.g., the current network round trip time for the current data center 104 to reach the client computing device) simply by measuring the time between when the server resource probe module transmits each server probe message over the network 108 to the client resource probe module 124, and when the server resource probe module receives a response to the server probe message from the client resource probe module. This measurement provides an estimate of the current network latency at zero payload. It will be appreciated that since each server probe message has a very tiny payload, the network 108 overhead associated with the server probe messages and their associated responses is negligible.
[045] Referring again to FIG. 1, in another embodiment of the disaster recovery technique described herein the client resource probe module 124 can periodically measure the current network latency associated with communicating with the current data center 104 and transmit this network latency measurement to the current data center. The client resource probe module can also periodically measure the current network bandwidth that is available to the client computing device 102 and transmit this network bandwidth measurement to the current data center. The client resource probe module can also periodically measure the current availability of one or more resources in the client computing device (including, but not limited to, its current battery availability (e.g., percent utilization), its current processor availability, and its current memory availability) and transmit these resource availability measurements to the current data center.
[046] Referring again to FIG. 1, the client resource probe module 124 can be
implemented in various ways using various mechanisms that are available on the client computing device 102. By way of example but not limitation, in one embodiment of the disaster recovery technique described herein the client resource probe module can periodically measure the current network latency associated with communicating with the current data center 104, and the current network bandwidth that is available to the client computing device, simply by measuring the time between when the application client 110 transmits each client request message over the network 108 to the application 114, and when the application client receives an application response to the client request message from the application. In another embodiment of the disaster recovery technique the client resource probe module can periodically transmit a client probe message over the network to the current data center, and then use a received probe response message from the current data center to determine the current network latency associated with
communicating with the current data center, and the current network bandwidth that is available to the client computing device. In yet another embodiment of the disaster recovery technique the client resource probe module can use one or more APIs that may exist in the operating system of the client computing device to determine the current availability of the just-described resources in the client computing device. It will be appreciated that there may be times when it may not be possible to measure certain resources in the client computing device (e.g., when the application client is a webpage that is running in a web browser, the application client may not be allowed to make certain resource measurements).
[047] Referring again to FIG. 1, in addition to the application 114, various additional program modules can be run in the current data center 104. As exemplified in FIG. 1, these additional program modules include, but are not limited to, a decision engine module 126, a current state store module 128, and a current server resource probe module 130. Exemplary functions of each of these additional program modules 126/128/130 are described in more detail hereafter. In an exemplary embodiment of the disaster recovery technique described herein, the decision engine and current state store modules 126 and 128 can be run either on the application server (not shown), or on another server computing device in the current data center 104. The current server resource probe module 130 is generally run on the application server.
[048] Generally speaking and referring again to FIG. 1, the current server resource probe module 130 operates cooperatively with both the client resource probe module 124 and a backup server resource probe module 134 that is running on each of the backup data centers 106/118 to periodically measure relevant network performance conditions and resource availability conditions at the application server in the current data center 104. The current server resource probe module 130 transmits these measurements to the decision engine module 126. The decision engine module 126 can use these measured conditions, along with other measured conditions described herein, to select an appropriate session state data replication option. More particularly and by way of example but not limitation, the current server resource probe module 130 can periodically measure the current network latency associated with communicating with the client computing device 102, and the current network bandwidth that is available to the client computing device. In an exemplary embodiment of the disaster recovery technique described herein these current network latency and current network bandwidth measurements can be made by the current server resource probe module 130 periodically transmitting a server probe message over the network 108 to the client computing device 102, and then using a received probe response message from the client computing device to determine the current network latency associated with communicating with the client computing device, and the current network bandwidth that is available to the client computing device. The current server resource probe module 130 can also periodically measure the current network latency associated with communicating with each of the backup data centers 106/118, and the current network bandwidth that is available to each of the backup data centers. In an exemplary embodiment of the disaster recovery technique these current network latency and current network bandwidth measurements can be made by the current server resource probe module 130 periodically transmitting a server probe message over the network 108 to each of the backup data centers 106/118, and then using a received probe response message from each of the backup data centers to determine the current network latency associated with communicating with each of the backup data centers, and the current network bandwidth that is available to each of the backup data centers.
[049] Referring again to FIG. 1, the current server resource probe module 130 can also periodically measure the current availability of one or more resources in the application server including, but not limited to, its current processor availability and its current memory availability. It will be appreciated that the current server resource probe module can make these resource availability measurements in various ways using various mechanisms that are available on the application server. By way of example but not limitation, the current server resource probe module can use one or more APIs that may exist in the operating system of the application server to determine the current availability of the just-described resources in the application server. The current server resource probe module can also include functionality that the client resource probe module 124 can use to measure the current network bandwidth that is available to the client computing device 102. By way of example but not limitation, upon receiving the aforementioned client probe message over the network 108 from the client resource probe module, the current server resource probe module can respond to the client probe message by transmitting a data payload over the network to the client resource probe module. The client resource probe module can measure the time it take to receive the data payload, and can then use this time and the size of the payload to calculate the current network bandwidth that is available to the client computing device.
[050] Referring again to FIG. 1, the decision engine module 126 periodically receives information over the network 108 from the client resource probe module 124, where this information includes the most recent network performance and resource availability measurements the client resource probe module has made at the client computing device 102. The decision engine module 126 also periodically receives information from the current server resource probe module 130, where this information includes the most recent performance and resource availability measurements the current server resource probe module has made at the application server in the current data center 104. The decision engine module 126 also receives the aforementioned various performance constraints and resource cost constraints, and a prioritization thereof, that can be specified by the application 114. Exemplary performance constraints that can be specified by the application include, but are not limited to, either a maximum allowable latency cost to the client computing device 102 associated with replicating the session state data, or a minimum allowable availability value for the application 114/132, or both this maximum allowable latency cost and this minimum allowable availability value. Exemplary resource costs and related constraints that can be specified by the application are described in more detail hereafter. The decision engine module 126 then selects an appropriate option for replicating the session state data for the interactive session 112 with the application 114 that increases the performance of the interactive session and/or decreases the resource costs for the interactive session. As will be described in more detail hereafter, this selection is made from a set of different session state data replication options each of which has different performance and resource cost tradeoffs, and can be based on the just- described constraints and their priorities. In other words, the decision engine module 126 determines how the session state data for the application 114 will be stored and replicated in the framework 100. The decision engine module 126 can use various methods to make this selection, examples of which are described in more detail hereafter.
[051] Referring again to FIG. 1, in an exemplary embodiment of the disaster recovery technique described herein, after an appropriate session state data replication option has been selected by the decision engine module 126 and implemented, the decision engine module can monitor the information it receives from the client and current server resource probe modules 124 and 130 on an ongoing basis. The decision engine module 126 can also monitor the various performance constraints and resource cost constraints, and the prioritization thereof, that can be specified by the application 114 on an ongoing basis. Whenever the decision engine module 126 detects changes in this received information, or in these constraints or their priorities, the decision engine module can again select an appropriate one of the session state data replication options that increases the performance of the interactive session 112 with the application 114 and/or decreases the resource costs for the interactive session, where this newly selected option can be either the same as or different than the previously selected option.
[052] Referring again to FIG. 1, the current state store module 128 stores the session state data for the interactive session 112 in the current data center 104 for the session state data replication options that involve storing the session state data in the current data center. In other words, the current state store module 128 inputs a data object (not shown) that includes the most recent session state data and stores it in the current data center 104. The current state store module 128 can also retrieve and output this stored data object whenever the application 114 requests it. The current state store module 128 thus ensures that the most recent session state data remains available while the interactive session 112 is active and as long as the current data center 104 is available. The current state store module 128 can store the data object in the current data center 104 in various ways. By way of example but not limitation, in one embodiment of the disaster recovery technique described herein the current state store module 128 can store the data object in either volatile memory or various types of non- volatile memory (e.g., magnetic disk storage, or the like) that may be available in the current data center 104. The current state store module 128 can also store the data object redundantly in the current data center 104 to protect against server failures there -within.
[053] As exemplified in FIG. 1, each of the backup data centers 106/118 hosts a copy of the application 132. In the event that the current data center 104 becomes unable to serve the interactive session 112 for whatever reason, one of the backup data centers (e.g., 106) will be automatically selected to take-over serving the interactive session. When this takeover occurs, in an exemplary embodiment of the disaster recovery technique described herein the copy of the application 132 will be run on one of the server computing devices in the selected backup data center 106 and the interactive session 112 will resume over the network 108 between this copy of the application and the application client 110. It is noted that until such a take-over occurs, the copy of the application 132 remains unused. The server computing device in the selected backup data center 106 that runs the copy of the application 132 and takes over serving the interactive session 112 is hereafter simply referred to as a backup application server.
[054] Referring again to FIG. 1, in addition to the copy of the application 132, various additional program modules can be run on each of the backup data centers 106/118. As exemplified in FIG. 1, these additional program modules include, but not limited to, the aforementioned backup server resource probe module 134 and a backup state store module 136. Exemplary functions of each of these additional program modules 134/136 are described in more detail hereafter. In an exemplary embodiment of the disaster recovery technique described herein, the backup state store module 136 is run either on the backup application server (not shown), or on another server computing device in the selected backup data center 106. The backup server resource probe module 134 is generally run on the backup application server.
[055] Referring again to FIG. 1, the backup state store module 136 receives the most recent session state data for the interactive session 112 from the current data center 104 and stores this received session state data in the backup data center 106 for the session state data replication options that involve storing the session state data in the backup data center. In other words, the backup state store module 136 inputs a data object (not shown) that includes the received session state data and stores it in the backup data center 106. The backup state store module 136 can also retrieve and output this stored data object whenever the copy of the application 132 requests it. The backup state store module 136 thus ensures that the session state data remains available while the interactive session 112 is active and as long as the backup data center 106 is available. The backup state store module 136 can store the data object in the backup data center 106 in various ways. By way of example but not limitation, in one embodiment of the disaster recovery technique described herein the backup state store module 136 can store the data object in either volatile memory or various types of non- volatile memory that may be available in the backup data center 106. In another embodiment of the disaster recovery technique the backup state store module 136 can write the data object to secondary storage such as a blob storage element which can be located either within the backup data center 106 or elsewhere in the cloud 138. This particular embodiment is advantageous in that it enhanced scalability and minimizes memory overhead. The backup state store module 136 can also store the data object redundantly in the backup data center 106 to protect against server failures there-within. It will be appreciated that the backup state store module 136 can store session state data for a plurality of interactive sessions that are served by the current data center 104.
[056] Generally speaking and referring again to FIG. 1, the backup server resource probe module 134 implements remote endpoint functionality that operates cooperatively with the current server resource probe module 130 and supports its aforementioned periodic measurement of the current network latency associated with communicating with the backup data center 106, and the current network bandwidth that is available to the backup data center 106. More particularly and by way of example but not limitation, the backup server resource probe module 134 can receive the server probe messages that are periodically sent by the current server resource probe module 130, and can immediately respond to each of these messages. Upon the current server resource probe module 130 receiving a probe response message from the backup server resource probe module 134, the current server resource probe module can determine the current network latency associated with communicating with the backup data center 106 by measuring the time between when each of the server probe messages is sent and when the corresponding probe response message is received (e.g., by measuring the current round trip network delay between the current data center 104 and the backup data center 106).
[057] Given the foregoing and referring again to FIG. 1, it will be appreciated that the replication of session state data that is provided for by the disaster recovery technique embodiments described herein ensures that the session state data for the application 114/132 remains available to the application client 110 (e.g., none of the session state data is lost) in the event that the current data center 104 goes offline for whatever reason, and thus becomes unable to continue serving the interactive session 112. As described heretofore, the disaster recovery technique embodiments replicate and thus protect the entire session state, including both user-entered data and server-computed data.
Accordingly, the disaster recovery technique embodiments allow the entire state of the interactive session that is taking place between the application client and the application to be maintained, even if the current data center goes offline and one of the backup data centers 106/118 takes over serving the interactive session. As such, if the current data center were to go offline for whatever reason, none of the session state data would have to be re-entered by the user or re-computed by the backup data centers.
[058] Although the disaster recovery technique embodiments described herein are discussed in the context of an interactive session being established between a given cloud application and a single client computing device that is being used by a single user, it is noted that the disaster recovery technique embodiments also support interactive sessions with cloud applications that include a plurality of users and a plurality of client computing devices which are jointly participating in the interactive session, and which may be located in different geographic regions. In other words, the disaster recovery technique embodiments support a situation where just a single application client is participating in an interactive session. The disaster recovery technique embodiments also support a situation where a plurality of application clients running on different client computing devices is collaboratively participating in a common interactive session at the same time (examples of such a situation include a plurality of geographically distributed users collaboratively playing a multi-player online game, or collaboratively participating in an online video teleconference, or collaboratively editing an online document, or collaboratively using a given enterprise application, among others). As will be appreciated from the more detailed description that follows, in such a multi-client interactive session the disaster recovery technique embodiments can select different session state data replication options for different ones of the application clients that are participating in the session depending on the individual network performance conditions and resource availability conditions at the client computing devices. Additionally, in such a multi-client interactive session any one or more of the application clients can leave at any time during the session, and the disaster recovery technique embodiments will ensure that the session continues correctly for the remaining application clients.
[059] Additionally, although the disaster recovery technique embodiments are discussed herein in the context of a given cloud application serving a single interactive session, it is noted that the disaster recovery technique embodiments also support a scenario where the application is serving a plurality of interactive sessions at the same time. In such a scenario the disaster recovery technique embodiments are implemented independently for each of the interactive sessions that is being served by the application.
1.2 Generic Libraries
[060] Generally speaking, this section describes various embodiments of the disaster recovery technique described herein where the aforementioned additional program modules that can be run on the client computing devices, the current data center, and the backup data centers can be implemented as generic libraries (not shown). The disaster recovery technique embodiments described in this section are hereafter simply referred to as generic libraries embodiments of the disaster recovery technique described herein. The generic libraries embodiments of the disaster recovery technique automatically resume an active interactive session with a cloud application at a backup data center and recover the most recent session state data in the event that the current data center goes offline for whatever reason. The generic libraries embodiments support both request-response based Hyper Text Transfer Protocol (HTTP) cloud applications and continuous interaction based web-sockets cloud applications. The generic libraries embodiments override the existing session state APIs in the cloud application and the existing communication APIs in its application client.
[061] In an exemplary implementation of the generic libraries embodiments of the disaster recovery technique described herein, each of the generic libraries sits just below the application layer in the network stack. Accordingly, interactions with the application that would normally occur through HTTP or web-sockets will occur through the generic libraries. In other words, the generic libraries intercept the HTTP and web-socket methods that applications use to communicate over the network, and add additional information (that is used to support the disaster recovery technique embodiments described herein) to application messages before they are transmitted over the network. Examples of such additional information include backup data center addresses and session state timestamps, among other types of information. In an exemplary implementation of the generic libraries embodiments this additional information is added to the application messages using JSON (JavaScript Object Notation) encoding. When the messages are received on the other end of the network, the additional information is removed by the generic libraries before the messages are passed to the application.
[062] As will be appreciated from the more detailed description that follows, the generic libraries embodiments of the disaster recovery technique are advantageous for various reasons including, but not limited to, the following. The generic libraries embodiments allow the additional program modules to be easily integrated with existing cloud applications without having to make significant code modifications thereto. The generic libraries embodiments also allow the cloud application developers to benefit from the functionality of the additional program modules without having to understand the details of their underlying implementation. The generic libraries embodiments also allow each of the additional program modules to operate in an application independent manner such that a given additional program module can be shared by different application clients or different cloud applications. The generic libraries embodiments thus eliminate the need for each different application that is hosted on the current data center to separately implement the decision engine module, the current state store module, and the current server resource probe module. The generic libraries embodiments also eliminate the need for each different copy of the application that is hosted on each backup data center to separately implement the backup server resource probe module and the backup state store module. The generic libraries embodiments also eliminate the need for each different application client that is run on each client computing device to separately implement the client state store module, the failure detector module, and the client resource probe module.
[063] More particularly and referring again to FIG. 1, the decision engine module 126 can be implemented as a generic library that is run on the current data center 104 and is shared by all of the different cloud applications (e.g., 114) that are hosted on the current data center. This particular generic library is hereafter simply referred to as a replication option selection library. All of the applications that are hosted on the current data center can input their performance constraints and resource cost constraints, and their prioritizations thereof, to the replication option selection library. For each of the applications that is hosted on the current data center, the replication option selection library can then use the application's constraints, and it's priorities thereof, to
independently select the optimal session state data replication option for each of the interactive sessions (e.g., 112) that either is currently taking place, or might take place in the future, with the application. It will be appreciated that implementing the decision engine module as a generic library is particularly well-suited to cloud computing scenarios, virtualized server scenarios, and other scenarios where a plurality of applications are being hosted on the same data center. In these scenarios the current data center can serve a large number of interactive sessions with a large number of applications, and the replication option selection library can be shared amongst all of the sessions and applications.
[064] Referring again to FIG. 1, the current state store and current server resource probe modules 128 and 130 can be implemented as another generic library that is run on the current data center 104 and is shared by the various cloud applications (e.g., 114) that are hosted on the current data center. The backup state store and backup server resource probe modules 136 and 134 can be implemented as yet another generic library that is run on each of the backup data centers 106/118 and is shared by the various cloud
applications. [065] Referring again to FIG. 1, the client state store, failure detector and client resource probe modules 120/122/124 can be implemented as yet another generic library that each application client 110 can include in its implementation. By way of example but not limitation, for web-browser-based applications that run as webpages on the client computing devices 102/1 16 this library can be implemented in the form of a JavaScript program (which is advantageous since JavaScript programs are supported by most web browsers). For other types of applications, this library can be implemented in various platform-specific forms (e.g., on client computing devices that run a Windows® (a registered trademark of Microsoft Corporation) operating system, this library might be implemented in the form of a .NET DLL (dynamic link library). It will be appreciated that both of these generic library implementations are advantageous in that they produce a generic library that is portable across a wide variety of client platforms.
1.3 Session State Data Replication Options
[066] Referring again to FIG. 1, this section provides a more detailed description of the different session state data replication options that can be used by the disaster recovery technique embodiments described herein. In other words, this section provides a more detailed description of different options for replicating the session state data for the interactive session 112 that is taking place over the network 108 between the application 114 and the application client 110. As stated heretofore and as will be further appreciated from the more detailed description that follows, each of the different session state data replication options has different performance and resource cost trade-offs. Generally speaking and as will be described in more detail hereafter, TCn is a cost parameter denoting a latency cost to the client computing device that is associated with a given replication option n. BCn is another cost parameter denoting a network bandwidth cost to the client computing device that is associated with the replication option n. BATCn is yet another cost parameter denoting a battery energy drain cost to the client computing device that is associated with the replication option n. BDn is yet another cost parameter denoting a network bandwidth cost to the current data center that is associated with the replication option n. As is appreciated in the art of cloud computing, cloud service providers often charge a given cloud application that they are hosting when data that is associated with the application leaves the data center that is hosting the application (such as when the current data center transmits session state data to a client computing device, or to a backup data center). SDn is yet another cost parameter denoting a data storage cost to the current data center that is associated with the replication option n. CDn is yet another cost parameter denoting a processing cost to the current data center that is associated with the replication option n.
[067] FIG. 2 summarizes the different session state data replication options that can be used by the disaster recovery technique embodiments described herein and the
performance and resource cost parameters (hereafter simply referred to as cost parameters) that are associated with each of the replication options. Each of the session state data replication options and its associated cost parameters are described in more detail hereafter.
[068] The first of the session state data replication options exemplified in FIG. 2 involves storing the most recent session state data for the interactive session on the client computing device. This particular option is hereafter simply referred to as replication option 1. Generally speaking and as will now be described in more detail, in replication option 1 the most recent session state data for the interactive session is added to each application response message that the application transmits over the network to the application client. Upon the application client receiving a given application response message, the application client will extract the most recent session state data from the application response message, and the client state store module will store this extracted data on the client computing device as described heretofore. The most recent session state data for the interactive session is also added to each of the client request messages that the application client transmits over the network to the application.
[069] In one embodiment of replication option 1 the application will add the entire set of most recent session state data to each application response message that the application transmits over the network to the application client. As such each application response message that the application client receives from the application will include the entire set of most recent session state data. The application client will also add the entire set of most recent session state data to each client request message that the application client transmits over the network to the application. As such, each client request message that the application receives from the application client will include the entire set of most recent session state data. Accordingly, in the event that the current data center goes offline and a given client request message from the application client is diverted to the backup data center, the copy of the application that receives the client request message will have the entire set of most recent session state data since it is included in the client request message. [070] In another embodiment of replication option 1 , rather than the entire set of most recent session state data being added to each application response message that the application transmits to the application client, the application will add the entire set of most recent session state data to the first application response message that the application it transmits to the application client, and to every Nth application response message thereafter that the application transmits to the application client (N being a first prescribed number greater than one). Similarly, rather than the entire set of most recent session state data being added to each client request message that the application client transmits to the application, the application client will add the entire set of most recent session state data to the first client request message that the application client transmits to the application, and to every Mth client request message thereafter that the application client transmits to the application (M being a second prescribed number greater than one). It will be appreciated that either N can be equal to M, or N can be greater than M, or N can be less than M. By way of example but not limitation, when N=M=3, the entire set of session state data will be added to the first application response message that the application transmits to the application client and every third application response message thereafter. Similarly, the entire set of session state data will be added to the first client request message that the application client transmits to the application and every third client request message thereafter.
[071] In yet another embodiment of replication option 1 , rather than the entire set of most recent session state data being added to each application response message that the application transmits to the application client, the application will add the entire set of most recent session state data to the first application response message that the application transmits to the application client, and the application will add just the delta in (e.g., just the modified portions of) the session state data to each application response message thereafter that the application transmits to the application client (e.g., the application will add to each application response message after the first response message just the changes in the session state data since it was last transmitted to the application client). Similarly, rather than the entire set of most recent session state data being added to each client request message that the application client transmits to the application, the application client will add the entire set of most recent session state data to the first client request message that the application client transmits to the application, and the application client will add just the delta in the session state data to each client request message thereafter that the application client transmits to the application (e.g., the application client will add to each client request message after the first client request message just the changes in the session state data since it was last transmitted to the application).
[072] In yet another embodiment of replication option 1 , rather than the entire set of most recent session state data being added to each application response message that the application transmits to the application client, the application will add the entire set of most recent session state data to the first application response message that the application transmits to the application client, and the application will add just the delta in the session state data to every Nth application response message thereafter that the application transmits to the application client. Similarly, rather than the entire set of most recent session state data being added to each client request message that the application client transmits to the application, the application client will add the entire set of most recent session state data to the first client request message that the application client transmits to the application, and the application client will add just the delta in the session state data to every Mth client request message thereafter that the application client transmits to the application. By way of example but not limitation, when N=M=3, the entire set of session state data will be added to the first application response message that the application transmits to the application client, and just the delta in the session state data will be added to every third application response message thereafter that the application transmits to the application client. Similarly, the entire set of session state data will be added to the first client request message that the application client transmits to the application, and just the delta in the session state data will be added to every third client request message thereafter that the application client transmits to the application.
[073] It will be appreciated that the most recent session state data can be added to the client request messages that the application client transmits over the network to the application, and the application response messages that the application transmits over the network to the application client, in various ways. By way of example but not limitation, in one embodiment of the disaster recovery technique described herein the most recent session state data can be added in the form of a cookie. In another embodiment of the disaster recovery technique the most recent session state data can be added into the body of each client request message and each application response message. An alternate embodiment of the disaster recovery technique is also possible where, rather than adding the most recent session state data to the client request and application response messages, the application client can transmit hidden messages that include the most recent session state data to the application, and the application can transmit similar hidden messages to the application client.
[074] Generally speaking, replication option 1 has certain performance and resource costs that in certain circumstances can make it undesirable compared to others of the session state data replication options exemplified in FIG. 2. More particularly, if the session state data is large (e.g., in the aforementioned example where a user is performing a flight search on a travel website, the session state data may include all of the available flights that meet the search criteria entered by the user thus far), then including the sessions state data in each request message that the application client transmits over the network to the application will slow down the client (and thus negatively impact the client's performance) since the network upload bandwidth that is available to most of the client computing devices will generally be significantly smaller than the network download bandwidth that is available thereto. Furthermore, including the session state data in each client request message that the application client transmits to the application and in each application response message that the application transmits to the application client may increase the request-response latency for the client (and thus negatively impact the client's performance).
[075] It is noted that the cost parameters associated with a given session state data replication option quantify the performance and/or resource utilization overheads associated with implementing the replication option compared to the performance and resource utilization when the replication option is not implemented (e.g., when the session state data is not replicated). For instance, it may be that the application already transmits all the session state data to the application client because the client can perform local processing on the session state data to serve some of the user's requests without having to contact the application. In this case the increase in request-response latency for the client would only be due to the client transmitting the session state data to the application.
Similar considerations are used when numerical values are assigned to the various cost parameters described in this section.
[076] TCI is a cost parameter denoting the latency cost to the client computing device associated with it receiving (e.g., downloading) the session state data from the application, and transmitting (e.g., uploading) the session state data to the application, in replication option 1. In other words, TCI denotes an increase in application client delay associated with this session state data reception and transmission. BC1 is another cost parameter denoting the network bandwidth cost to the client computing device associated with it receiving the session state data from the application, and transmitting the session state data to the application, in replication option 1. In other words, BC1 denotes the amount the application client has to pay for this session state data reception and transmission. BC1 may be zero when the client computing device is connected to the network via a wired connection having unlimited data. Alternatively, BC1 may be very high when the client computing device is connected to the network via a conventional cellular data connection operating in roaming mode. BATC1 is yet another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client receiving the session state data from the application, and transmitting the session state data to the application, in replication option 1. In other words, B ATC 1 denotes how much battery energy in the client computing device is used for this session state data reception and transmission. While BATC1 may not be very relevant when the client computing device is operating from an AC (alternating current) power source, B ATC 1 is quite relevant when the client computing device is operating from battery power (e.g., when the client computing device is either a smartphone, or a tablet computer, or the like).
[077] The second of the session state data replication options exemplified in FIG. 2 involves synchronously storing the most recent session state data for the interactive session in one or more of the backup data centers. This particular option is hereafter simply referred to as replication option 2. In replication option 2 the current data center uploads (e.g., transmits) the most recent session state data over the network to one or more of the backup data centers whenever the session state changes, thus ensuring that the backup data centers are storing the most recent session state data. The backup state store module will store this most recent session state data in the one or more backup data centers as described heretofore, and will then transmit a confirmation message to the current data center indicating that the most recent session state data has been successfully stored in the backup data centers. By way of example but not limitation, whenever the application receives a client request message from the application client, the current data center will upload the most recent session state data over the network to one or more of the backup data centers. The application will then wait to respond to the client request message until the current data center receives the confirmation message from at least one of the backup data centers. In one embodiment of replication option 2, the current data center will upload the entire set of most recent session state data to one or more of the backup data centers whenever the session state changes. In another embodiment of replication option 2, the current data center will upload the entire set of most recent session state data to one or more of the backup data centers the first time the session state changes, and the current data center will upload just the delta in the session state data to the backup data centers thereafter (e.g., the current data center will upload just the changes in the session state data since it was last uploaded to the backup data centers).
[078] Replication option 2 has the following performance and resource cost trade-offs. Since replication option 2 does not burden the client computing device, it avoids any network bandwidth cost thereon. However, replication option 2 does increase the time that the application client has to wait to receive application response messages from the application, which negatively impacts the client's performance. TC2 is a cost parameter denoting the latency cost to the client computing device associated with the application client having to wait to receive application response messages from the application in replication option 2. In other words, TC2 denotes an increase in application client delay associated with this having to wait. Replication option 2 also increases the cost to the current data center for hosting the application since the current data center has to pay for the network bandwidth it uses to upload the most recent session state data to one or more of the backup data centers. BD2 is another cost parameter denoting the network bandwidth cost to the current data center associated with it uploading the session state data to one or more of the backup data centers in replication option 2. In other words, BD2 denotes the amount the current data center has to pay for this session state data upload. Replication option 2 also presents the current data center with the cost of storing the session state data in one or more of the backup data centers. SD2 is yet another cost parameter denoting the data storage cost to the current data center associated with it storing the session state data in one or more of the backup data centers in replication option 2.
[079] The third of the session state data replication options exemplified in FIG. 2 involves storing an encoded version of the most recent session state data for the interactive session on the client computing device, where the encoding reduces the data size. This particular option is hereafter simply referred to as replication option 3. Generally speaking, in replication option 3 an encoded version of the most recent session state data for the interactive session is added to each application response message that the application transmits over the network to the application client, and an encoded version of the most recent session state data for the interactive session is added to each of the client request messages that the application client transmits over the network to the application. As such each of the application response messages that the application client receives from the application will include an encoded version of the most recent session state data. Upon the application client receiving a given application response message, the application client will extract the encoded version of the most recent session state data from the application response message, and the client state store module will store this extracted data on the client computing device as described heretofore. Similarly, each client request message that the application receives from the application client will include an encoded version of the most recent session state data. Accordingly, in the event that the current data center goes offline and a given client request message from the application client is diverted to the backup data center, the copy of the application that receives the client request message will have an encoded version of the most recent session state data since it is included in the client request message. It will be appreciated that an encoded version of the most recent session state data can be added to the client request messages and the application response messages in various ways including, but not limited to, the aforementioned various ways that were described in relation to replication option 1.
[080] As will now be described in more detail, the encoded version of the session state data that is transmitted over the network and stored on the client computing device in replication option 3 is smaller in size than the non-encoded version of the session state data that is transmitted over the network and stored on the client computing device in replication option 1. In fact, for certain applications the encoded version of the session state data may be substantially smaller than the non-encoded version thereof. More particularly and by way of example but not limitation, in one embodiment of the disaster recovery technique described herein the encoded version of the most recent session state data includes just the data that the user most recently entered into the client computing device (hereafter simply referred to as most recent user-entered data). In the event that the current data center goes offline and a given client request message from the application client is diverted to a backup data center, the copy of the application that receives the client request message can process the most recent user-entered data and re-create the remainder of the most recent session state data. In another embodiment of the disaster recovery technique the encoded version of the session state data is a compressed version of all of the session state data that is transmitted and stored in replication option 1. This compression and the subsequent decompression of the session state data can be performed using any data compression method. By way of example but not limitation, in an exemplary embodiment of the disaster recovery technique a conventional delta compression method is used for this compression and decompression; this particular method is advantageous since, as described heretofore, the session state can change incrementally as user interactions with the application progress.
[081] TC3 is a cost parameter denoting the latency cost to the client computing device associated with it receiving the encoded session state data from the application, and transmitting the encoded session state data to the application, in replication option 3. In other words, TC3 denotes an increase in application client delay associated with this encoded session state data reception and transmission. BC3 is another cost parameter denoting the network bandwidth cost to the client computing device associated with it receiving the encoded session state data from the application, and transmitting the encoded session state data to the application, in replication option 3. In other words, BC3 denotes the amount the application client has to pay for this encoded session state data reception and transmission. BC3 may be zero when the client computing device is connected to the network via a wired connection having unlimited data. Alternatively, BC3 may be very high when the client computing device is connected to the network via a cellular data connection operating in roaming mode. BATC3 is yet another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client receiving the encoded session state data from the application, and transmitting the encoded session state data to the application, in replication option 3. In other words, BATC3 denotes how much battery energy in the client computing device is used for this encoded session state data reception and transmission. While BATC3 may not be very relevant when the client computing device is operating from an AC power source, BATC3 is quite relevant when the client computing device is operating from battery power.
[082] Since replication option 3 reduces the size of the session state data that is transmitted over the network and stored on the client computing device, TC3 will be less than TCI, BC3 will be less than BC1, and BATC3 will be less than BATC1. In fact, for certain applications TC3, BC3 and BATC3 may be substantially less than TCI, BC1 and BATC1 respectively. CD3 is yet another cost parameter denoting the processing cost to the current data center associated with the application decoding the encoded session state data in replication option 3. It will be appreciated that the application in the current data center has to decode the encoded session state data in each client request message it receives from the client (e.g., either process the user-entered data and re-create the remainder of the most recent session state data, or decompress the compressed session state data to produce the most recent session state data) in order to be able to respond to the client request message. This decoding operation will increase the processor overhead in the current data center, and will also slightly increase the request-response latency for the client.
[083] The fourth of the session state data replication options exemplified in FIG. 2, hereafter simply referred to as replication option 4, also involves storing the most recent session state data for the interactive session on the client computing device. Replication option 4 is similar to replication option 1 in that the most recent session state data for the interactive session is added to each application response message that the application transmits over the network to the application client. Upon the application client receiving a given application response message, the application client will extract the most recent session state data from the application response message, and the client state store module will store this extracted data on the client computing device as described heretofore. It is noted that the most recent session state data can be added to the application response messages in various ways including, but not limited to, the aforementioned various ways that were described in relation to replication option 1.
[084] Replication option 4 differs from replication option 1 in the following ways. In replication option 4, rather than the most recent session state data for the interactive session being added to each of the client request messages that the application client transmits over the network to the application, failure detection software on the client computing device (e.g., the aforementioned failure detector module) is responsible for detecting when the current data center that is serving the interactive session has gone offline and thus has become unable to continue serving the session. Whenever this failure detection software detects that the current data center has gone offline and become unable to continue serve the interactive session, the application client will upload the most recent session state data for the interactive session to a backup data center that has been selected to take over serving the interactive session. The most recent session state data for the interactive session will then be added to each application response message that the copy of the application transmits over the network to the application client. Upon the application client receiving a given application response message from the copy of the application, the application client will extract the most recent session state data from the application response message, and the client state store module will store this extracted data on the client computing device as described heretofore.
[085] TC4 is a cost parameter denoting the latency cost to the client computing device associated with it receiving the session state data from the application, and transmitting the session state data to a backup data center whenever the current data center becomes unable to continue serving the interactive session, in replication option 4. In other words, TC4 denotes an increase in application client delay associated with this session state data reception and transmission. BC4 is another cost parameter denoting the network bandwidth cost to the client computing device associated with it receiving the session state data from the application, and transmitting the session state data to a backup data center whenever the current data center becomes unable to continue serving the interactive session, in replication option 4. In other words, BC4 denotes the amount the application client has to pay for this session state data reception and transmission. BC4 may be zero when the client computing device is connected to the network via a wired connection having unlimited data. Alternatively, BC4 may be very high when the client computing device is connected to the network via a cellular data connection operating in roaming mode. BATC4 is yet another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client receiving the session state data from the application, and the client transmitting the session state data to a backup data center whenever the current data center becomes unable to continue serving the interactive session, in replication option 4. In other words, BATC4 denotes how much battery energy in the client computing device is used for this session state data reception and
transmission. While BATC4 may not be very relevant when the client computing device is operating from an AC power source, BATC4 is quite relevant when the client computing device is operating from battery power. Since replication option 4 does not involve the application client adding the session state data to each client request message the application client transmits over the network to the application (as is done in replication option 1), TC4 will be less than TCI, BC4 will be less than BC1, and BATC4 will be less than BATC1.
[086] The fifth of the session state data replication options exemplified in FIG. 2 involves asynchronously storing the most recent session state data for the interactive session in one or more of the backup data centers. This particular option is hereafter simply referred to as replication option 5. In replication option 5 the current data center uploads the most recent session state data (either the entire set thereof, or just the delta therein, as described in relation to replication option 2) over the network to one or more of the backup data centers whenever the session state changes, thus ensuring that the backup data centers are storing the most recent session state data. The backup state store module will store this most recent session state data in the one or more backup data centers as described heretofore, and will then transmit a confirmation message to the current data center indicating that the most recent session state data has been successfully stored in the backup data centers. By way of example but not limitation, whenever the application receives a client request message from the application client, the current data center will upload the most recent session state data over the network to one or more of the backup data centers.
[087] However, rather than the application waiting to respond to the client request message until the current data center receives the confirmation message from at least one of the backup data centers (as is done in replication option 2), in replication option 5 the application will respond to the client request message before the current data center receives this confirmation message from at least one of the backup data centers, but the application client will wait to expose the application response to the user of the client computing device until the confirmation message is received from at least one of the backup data centers. Accordingly, in replication option 5, during the time that the most recent session state data is being uploaded to the backup data centers the client computing device is already downloading the application response, thus hiding some of the network round trip time between the current and backup data centers. It will be appreciated that doing this introduces a risk that if the current data center were to fail after the most recent session state data is transmitted to the backup data centers, and the backup data centers fail to receive the most recent session state data due to a network data transmission error, then the most recent session state data will not be available to the copies of the application when one of the backup data centers is selected to take over serving the interactive session.
[088] Replication option 5 can address this risk (e.g., prevent a failure of the current data center from causing the most recent session state data to be lost) in various ways. By way of example but not limitation, in one embodiment of the disaster recovery technique described herein, the application can wait a prescribed period of time after the most recent session state data for the interactive session is uploaded to one or more backup data centers before the application responds to the client request message, where this period of time is made large enough to ensure that the most recent session state data has left the current data center (e.g., has left the current data center's internal network (not shown) and has been transmitted onto the external network). In another embodiment of the disaster recovery technique the current data center can add forward error correction (FEC) codes to the most recent session state data before it is uploaded to the backup data centers. As is appreciated in the art of data communication, if just some of the uploaded data is successfully received by a given backup data center and the rest of the uploaded data is lost by the either the current data center's internal network or the external network, the FEC codes allow the backup data center to reconstruct the data that was lost.
[089] Replication option 5 has the following performance and resource cost trade-offs. Since replication option 5 does not burden the client computing device, it avoids any network bandwidth cost thereon. However, replication option 5 does very slightly increase the time that the application client has to wait to receive application response messages from the application, which has a very small negative impact on the client's performance. TC5 is a cost parameter denoting the latency cost to the client computing device associated with the application client having to wait to receive application response messages from the application in replication option 5. In other words, TC5 denotes an increase in application client delay associated with the extra amount of time that the application client has to wait to receive application response messages from the application in replication option 5. BATC5 is another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client having to wait this extra amount of time. Given the foregoing, it will be appreciated that TC5 will generally be less than TC2.
[090] Replication option 5 increases the cost to the current data center for hosting the application since the current data center has to pay for the network bandwidth it uses to upload the most recent session state data to one or more of the backup data centers. BD5 is yet another cost parameter denoting the network bandwidth cost to the current data center associated with it uploading the session state data and FEC codes to one or more of the backup data centers in replication option 5. In other words, BD5 denotes the amount the current data center has to pay for this session state data and FEC codes upload. Due to the addition of the FEC codes to the data that is transmitted, BD5 will be greater than BD2. Replication option 5 also presents the current data center with the cost of storing the session state data in one or more of the backup data centers. SD5 is yet another cost parameter denoting the data storage cost to the current data center associated with it storing the session state data in one or more of the backup data centers in replication option 5. It will be appreciated that SD5 will be the same as SD2.
[091] The sixth of the session state data replication options exemplified in FIG. 2 is a variant of replication option 3 that involves storing an encrypted version of the most recent session state data for the interactive session on the client computing device. This particular option is hereafter simply referred to as replication option 6. Replication option 6 is applicable to a situation where the session state data that is computed by the application is to be stored on the client computing device but is not to be disclosed to the user of the client computing device. One example of such a situation is an online card game application where the session state data that is computed by the application includes the hand of cards that is generated for all of the users that are currently playing the game. Encrypting this session state data prevents the hand of cards for users other than the client computing device's user, and the hand of cards for the computerized player, from being disclosed to the client computing device's user. The session state data can be encrypted using any conventional data encryption method.
[092] TC6 is a cost parameter denoting the latency cost to the client computing device associated with it receiving the encrypted session state data from the application in replication option 6. In other words, TC6 denotes an increase in application client delay associated with this encrypted session state data reception. BC6 is another cost parameter denoting the network bandwidth cost to the client computing device associated with it receiving the encrypted session state data from the application in replication option 6. In other words, BC6 denotes the amount the application client has to pay for this encrypted session state data reception. BC6 may be zero when the client computing device is connected to the network via a wired connection having unlimited data. Alternatively, BC6 may be very high when the client computing device is connected to the network via a cellular data connection operating in roaming mode. BATC6 is yet another cost parameter denoting the battery energy drain cost to the client computing device associated with the application client receiving the encrypted session state data from the application in replication option 6. In other words, BATC6 denotes how much battery energy in the client computing device is used for this encrypted session state data reception. While BATC6 may not be very relevant when the client computing device is operating from an AC power source, BATC6 is quite relevant when the client computing device is operating from battery power. CD6 is yet another cost parameter denoting the processing cost to the current data center associated with the application encrypting the encoded session state data in replication option 6. This encryption operation will increase the processor overhead in the current data center, and will also slightly increase the request-response latency for the client.
[093] The seventh of the session state data replication options exemplified in FIG. 2, hereafter simply referred to as replication option 7, is a hybrid option that can include any combination of replication options 1-6. By way of example but not limitation, in the aforementioned situation where a plurality of application clients running on different client computing devices is jointly participating in an interactive session at the same time, a session state data replication option can be independently selected for each of the client computing devices. In other words and referring again to FIG. 1, the decision engine module may decide to store the session state data on all of the client computing devices. The decision engine module may also decide to store the session state data on just some of the client computing devices and in one or more of the backup data centers. The decision engine module may also decide to store the session state data on just one or more of the backup data centers.
[094] Additionally, replication option 2 can be used to store the session state data for the interactive session in a backup data center that is near the current data center, and replication option 5 can be used to store the session state data in another backup data center that is not near (e.g., is far away from) the current data center. This combined utilization of replication options 2 and 5 is advantageous since the backup data center that is near the current data center is not likely to fail at the same time as the backup data center that is not near the current data center (especially in the situation where the failure is triggered by an earthquake, or fire, or lightning strike, or other type of natural disaster). Replication option 3 can also be combined with either replication option 2 or replication option 5 so that an encoded version of the session state data for the interactive session is either synchronously or asynchronously stored in each of the backup data centers. As indicated in FIG. 2, various combinations of the different performance and resource cost parameters that are associated with replication options 1-6 become relevant in replication option 7.
[095] The different performance and resource cost parameters that are associated with replication options 1-7 might not be static and their values might not be known upfront when a given cloud application is being developed or deployed. By way of example but not limitation, for an interactive session involving a user who is using a client computing device that is operating from battery power and is connected to the network via a wireless connection such as a conventional cellular data connection, or WiFi connection, or the like (e.g., when the client computing device is either a smartphone, or a tablet computer, or the like), the client's network bandwidth cost (e.g., BC1) and battery energy drain cost (e.g., BATC1) may be very high. On the other hand, for an interactive session involving a user who is using a client computing device that is operating from AC power and is connected to the network via a wired connection, the client's network bandwidth cost and battery energy drain cost may be negligible. Furthermore, the current data center's network bandwidth cost (e.g., BD2) and data storage cost (e.g., SD2) may change over time depending on various factors such as changes in the load on the current and backup data centers. Yet furthermore, it will be appreciated that the time the application client has to wait to receive application response messages from the application (e.g., TC2) depends on how far the backup data centers are from the current data center. When a given cloud application is first deployed, this client wait time may be high since the application may initially be hosted on just two data centers that are located far apart from one another. As the usage of the application grows, the application may be deployed in additional data centers in different locations, or the application may start to rent data center capacity offered by cloud service providers in different locations, either of which may decrease the distance between the current data center and one or more of the backup data centers, thus decreasing the client wait time (e.g., decreasing TC2).
[096] It will thus be appreciated that the values of the different performance and resource cost parameters that are associated with replication options 1-7 can be different for different interactive sessions with a given cloud application, and can change over time. The disaster recovery technique embodiments described herein can select an appropriate session state data replication option when the application is first deployed since some of the performance and resource cost parameters become known at this time (e.g., the initial number of data centers hosting the application becomes known, the physical locations and thus the distances between the various pairs of the data centers becomes known, and the network bandwidth and data storage costs for each of the data centers become known). The disaster recovery technique embodiments can also select an appropriate session state data replication option whenever there are configuration changes in the computing and networking infrastructure that supports the application (e.g., whenever the number and/or physical locations of data centers hosting the application changes, or whenever the network bandwidth and data storage costs for one or more of the data centers changes).
[097] The disaster recovery technique embodiments described herein can also select an appropriate session state data replication option whenever a new interactive session is established between a given cloud application that is running on a given data center and an application client that is running on a given client computing device that is being used by a given user (in this case, all of the performance and resource cost parameters that are associated with replication options 1-7 can be considered). After the interactive session is established and an appropriate session state data replication option has been selected, the disaster recovery technique embodiments can also monitor all of the performance and resource cost parameters and again select an appropriate session state data replication option whenever there are changes in these parameters, where this newly selected option can be either the same as or different than the previously selected option.
1.4 Selecting an Option for Replicating Session State Data
[098] This section provides a more detailed description of the aforementioned various methods that the decision engine module can use to select an appropriate option for replicating the session state data for the interactive cloud application that is being hosted on the current data center. This selection is made from a set of different session state data replication options that includes, but is not limited to, the aforementioned replication options 1-7. As described heretofore, each of these replication options has different performance and resource cost trade-offs. As also described heretofore and will now be described in more detail, this selection can be based on various network performance conditions and resource availability conditions that can be measured by the client resource probe module, current server resource probe module, and backup server resource probe module. This selection can also be based on various performance constraints and resource cost constraints, and a prioritization thereof, that can be specified by the application.
[099] FIG. 3 illustrates one embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application that is being hosted on the current data center. As exemplified in FIG. 3, the process starts in block 300 with estimating the latency cost to a client computing device associated with each of the session state data replication options in the set of different session state data replication options. In other words, and by way of example but not limitation, TCI, TC2, TC3, TC4, TC5 and TC6 can be estimated using the various measurement techniques described heretofore. The session state data replication option having the smallest estimated latency cost to the client computing device is then selected as the option to be used for replicating the session state data for an interactive session between the application and the client (block 302). It will thus be appreciated that the process embodiment exemplified in FIG. 3 minimizes the latency cost to the client computing device associated with replicating the session state data, and thus minimizes the delay overhead that a user will experience due to the session state data replication.
[0100] FIG. 4 illustrates another embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application that is being hosted on the current data center. As exemplified in FIG. 4, the process starts in block 400 with inputting one or more performance constraints that are specified by the application. These performance constraints include a maximum allowable latency cost to a client computing device associated with replicating the session state data for the interactive session (e.g., a constraint on TCI, TC2, TC3, TC4, TC5 and TC6). The performance constraints can optionally also include a minimum allowable availability for the application. One or more resource cost constraints that are specified by the application are then input (block 402). These resource cost constraints can include one or more of client resource cost constraints or data center resource cost constraints. Exemplary client resource cost constraints include, but are not limited to, one or more of a constraint on the network bandwidth cost to the client computing device associated with replicating the session state data for the interactive session (e.g., a constraint on BC1, BC3, BC4 and BC6), or a constraint on the battery energy drain cost to the client computing device associated with replicating this session state data (e.g., a constraint on BATC1, BATC3, BATC4, BATC5 and BATC6), or a constraint on the processing cost to (e.g., a processor utilization budget for) the client computing device associated with replicating this session state data, or a constraint on the memory cost to (e.g., a memory utilization budget for) the client computing device associated with replicating this session state data. Exemplary data center resource cost constraints include, but are not limited to, one or more of a constraint on the network bandwidth cost to the current data center associated with replicating the session state data for the interactive session (e.g., a constraint on BD2 and BD5), or a constraint on the data storage cost to the current data center associated with replicating this session state data (e.g., a constraint on SD2 and SD5), or a constraint on the processing cost to the current data center associated with replicating this session state data.
[0101] Referring again to FIG. 4, after the performance and resource cost constraints have been input (blocks 400 and 402), the latency cost to the client computing device associated with each of the session state data replication options in the set of different session state data replication options is estimated (block 404) as described heretofore. The session state data replication options whose estimated latency cost to the client computing device is less than or equal to the just-described maximum allowable latency cost to the client computing device are then identified (block 406). One of these identified options that meets the resource cost constraints is then selected as the option to be used for replicating the session state data for the interactive session (block 408). It will thus be appreciated that the process embodiment exemplified in FIG. 4 can jointly minimize both the latency cost to the client computing device associated with replicating the session state data, and one or more prescribed resource costs that are associated with replicating the session state data.
[0102] FIG. 5 illustrates yet another embodiment, in simplified form, of a process for selecting an appropriate option for replicating the session state data for the cloud application that is being hosted on the current data center. As exemplified in FIG. 5, the process starts in block 500 with inputting one or more resource cost constraints that are specified by the application. The latency cost to a client computing device associated with each of the session state data replication options in the set of different session state data replication options is then estimated (block 502). The session state data replication options that meet the resource cost constraints are then identified (block 504). The one of these identified options that has the smallest estimated latency cost to the client computing device is then selected as the option to be used for replicating the session state data for the interactive session (block 506). It will thus be appreciated that the process embodiment exemplified in FIG. 5 can also jointly minimize both the latency cost to the client computing device associated with replicating the session state data, and one or more prescribed resource costs that are associated with replicating the session state data.
[0103] It is noted that the resource cost constraints can be specified by the application in various ways. By way of example but not limitation, the resource cost constraints can be specified for individual interactive sessions with the application. The resource cost constraints can also be specified as a total resource cost budget that is to be allocated across a plurality of interactive sessions with the application. In this case, the decision engine module can allocate the total resource cost budget in various ways including, but not limited to, the following. The decision engine module can allocate the total resource cost budget on the fly to individual interactive sessions having a current network latency that is higher than a prescribed threshold (e.g., additional resources can be used to reduce the network latency and thus reduce the user response times). The decision engine module can also allocate the total resource cost budget over a prescribed period of time. More particularly and by way of example but not limitation, specific amounts of the budget can be allocated to specific days of the week. The budget can be allocated in proportion to the application load (e.g., the user demand for the application) that is expected over the period of time. In other words, different portions of the budget can be allocated to different segments of the period of time based on the application load that is expected during each of the segments. [0104] An alternate embodiment of the disaster recovery technique described herein is also possible where the one or more performance constraints that are specified by the application, and the one or more resource cost constraints that are specified by the application, can be combined into a unified constraint by applying a weight to each of the performance constraints and each of the resource cost constraints. It will be appreciated that the weights that are applied to the different performance and resource cost constraints serve to prioritize these constraints. By way of example but not limitation, these weights can be used to specify the relative importance to the application of the latency cost to the client computing device as compared to the network bandwidth cost to the client computing device. The weights can be determined in various ways. By way of example but not limitation, in an exemplary implementation of this alternate embodiment the weights are pre-determined by the application.
1.5 Process Framework
[0105] FIG. 6 illustrates an exemplary embodiment, in simplified form, of the server side of a process for ensuring the availability of a cloud application that is being hosted on a current data center. As exemplified in FIG. 6, the process starts in block 600 with selecting an option for replicating session state data for the application. As described heretofore, this selection is made from a set of different session state data replication options each of which has different performance and resource cost trade-offs, and the selected option determines how the session state data for the application is to be replicated. The selected option is then implemented (block 602), where the implementation results in the session state data for the application being replicated outside of the current data center, thus ensuring that this data remains available in the event that that current data center goes offline.
[0106] FIG. 7 illustrates an exemplary embodiment, in simplified form, of the client side of a process for providing disaster recovery for the cloud application that is being hosted on the current data center. As exemplified in FIG. 7, the process starts in block 700 with receiving a message from the current data center specifying an option for replicating session state data for the application that has been selected for a given client computing device. The specified option is then implemented (block 702), where the implementation results in the session state data for the application being replicated outside of the current data center. 2.0 Provisioning and Using Spare Data Center Capacity
[0107] Referring again to FIG. 1, the disaster recovery technique embodiments described heretofore assume that a plurality of data centers (e.g., 104/106/118) exist in the cloud 138, and also assume that this plurality of data centers includes one or more backup data centers 106/118 which have sufficient spare (e.g., idle) capacity (e.g., sufficient spare/idle processing capacity and sufficient spare/idle data storage capacity) to take over the current application load of the current data center 104 if it fails. In other words, the disaster recovery technique embodiments described heretofore assume that the backup data centers have sufficient spare capacity to take over serving the interactive sessions 112 that are currently taking place with the cloud applications 114 that are currently being hosted by the current data center in the event that it fails.
[0108] This section describes an exemplary embodiment of the disaster recovery technique described herein that adds spare capacity to each of the data centers in the cloud, where this spare capacity is sufficient to allow a failure of any one of the data centers to be accommodated by the remaining data centers. In other words, if any one of the data centers in the cloud fails, the spare capacity that is added to each of the remaining data centers by the disaster recovery technique embodiment described in this section allows the current application load of the failed data center to be distributed across the remaining data centers (e.g., the remaining data centers can collectively take over serving the interactive sessions that were being served by the failed data center).
[0109] FIG. 8 illustrates an exemplary embodiment, in simplified form, of a process for provisioning and using spare data center capacity in the cloud. As exemplified in FIG. 8, the process starts in block 800 with determining an amount of spare server capacity (e.g., the number of additional servers) that has to be added to each of the data centers in the cloud in order for the cloud to be able to take over the current application load of any one of the data centers if it fails. This determination is made by taking into account different factors which include, but are not limited to, one or more of the following: the total number of data centers in the cloud, the current server capacity of (e.g., the current number of servers in) each of the data centers, and the geographic location of each of the data centers. After the amount of spare server capacity to be added to each of the data centers has been determined (block 800), it is then directed that this determined amount of spare server capacity be added to each of the data centers (block 802). Then, whenever any one of the data centers fails, the current application load of the failed data center will be distributed across the spare server capacity that was added to each of the remaining data centers (804).
[0110] It is noted that the process embodiment exemplified in FIG. 8 is applicable to situations where either the current server capacity of each of the data centers in the cloud in the cloud is either the same, or the current server capacity of one or more of the data centers is different than the current server capacity of others of the data centers. The process embodiment exemplified in FIG. 8 is also applicable to situations where either all of the data centers are located in the same geographic region, or the data centers are located in two or more different geographic regions, where these different geographic regions may be in either the same time zone or different time zones. It is also noted that depending on the particular factors that are taken into account when the just-described determination of spare server capacity is made, either the same amount of spare server capacity can be added to each of the data centers, or different amounts of spare server capacity can be added to different ones of the data centers.
[0111] Given the foregoing, it will be appreciated that the process embodiment exemplified in FIG. 8 is advantageous in that it can minimize the amount of spare capacity that has to be added to each of the data centers in the cloud in order for the cloud to be able to take over the current application load of any one of the data centers if it fails. Accordingly, the process embodiment exemplified in FIG. 8 can minimize the expense associated with provisioning spare data center capacity.
3.0 Additional Embodiments
[0112] While the disaster recovery technique has been described by specific reference to embodiments thereof, it is understood that variations and modifications thereof can be made without departing from the true spirit and scope of the disaster recovery technique. By way of example but not limitation, an embodiment of the disaster recovery technique described herein is possible where the session state data for a given interactive session between a given cloud application and a given application client includes one or more data items that cannot be stored on just the client computing device that is running the application client. By way of example but not limitation, when the interactive session is a Hyper Text Transfer Protocol Secure (HTTPS) session (such as when a user makes ecommerce transactions on a secure banking website), the cryptographic keys associated with the HTTPS session have to be stored on both the application server and the client computing device, and the application and application client already take care of this. In such a situation, just session state data replication options that store the most recent session state data on one or more backup data centers (e.g., either the aforementioned replication option 2 or replication option 5) will be used since the backup data centers have to have the private (e.g., secret) key to even decrypt any client request messages the backup data centers may receive from the application client in the event that the current data center goes offline.
[0113] Another embodiment of the disaster recovery technique described herein is also possible where the most recent session state data that is stored on a given client computing device during a given interactive session is encrypted by the client state store module. More particularly, after the application client has extracted the most recent session state data for the interactive session from a given application response message, the client state store module will encrypt this extracted data using any conventional data encryption method, and will then store this encrypted data and an associated encryption key on the client computing device. Additionally, in replication option 4, whenever the failure detector module detects that the current data center has gone offline, the session resume request message that the application client transmits to the backup data center that has been selected to take over serving the interactive session will include both the encrypted most recent session state data for the interactive session and the associated encryption key that are stored on the client computing device. This allows the backup data center to decrypt the encrypted most recent session state data before it resumes the interactive session with the client computing device. This particular embodiment is applicable in situations where the session state data includes proprietary information that may not be appropriate to expose to a user. Examples of such proprietary information include customer profiling data in an customer database, a selection of advertisements to be shown or discount codes to be offered during the session, certain types of information generated by an ecommerce application, and the like. Additionally, in an online card game application the session state data may include the hand of cards that is generated for all of the users that are currently playing the game. Storing such proprietary information in encrypted form on the client computing device is advantageous since the information is kept secure.
[0114] It is also noted that any or all of the aforementioned embodiments can be used in any combination desired to form additional hybrid embodiments. Although the disaster recovery technique embodiments have been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described heretofore. Rather, the specific features and acts described heretofore are disclosed as example forms of implementing the claims.
4.0 Exemplary Operating Environments
[0115] The disaster recovery technique embodiments described herein are operational within numerous types of general purpose or special purpose computing system environments or configurations. FIG. 9 illustrates a simplified example of a general- purpose computer system on which various embodiments and elements of the disaster recovery technique, as described herein, may be implemented. It is noted that any boxes that are represented by broken or dashed lines in the simplified computing device 900 shown in FIG. 9 represent alternate embodiments of the simplified computing device. As described below, any or all of these alternate embodiments may be used in combination with other alternate embodiments that are described throughout this document. The simplified computing device 900 is typically found in devices having at least some minimum computational capability such as personal computers (PCs), server computers, handheld computing devices, laptop or mobile computers, communications devices such as cell phones and personal digital assistants (PDAs), multiprocessor systems, microprocessor-based systems, set top boxes, programmable consumer electronics, network PCs, minicomputers, mainframe computers, and audio or video media players.
[0116] To allow a device to implement the disaster recovery technique embodiments described herein, the device should have a sufficient computational capability and system memory to enable basic computational operations. In particular, the computational capability of the simplified computing device 900 shown in FIG. 9 is generally illustrated by one or more processing unit(s) 910, and may also include one or more graphics processing units (GPUs) 915, either or both in communication with system memory 920. Note that that the processing unit(s) 910 of the simplified computing device 900 may be specialized microprocessors (such as a digital signal processor (DSP), a very long instruction word (VLIW) processor, a field-programmable gate array (FPGA), or other micro-controller) or can be conventional central processing units (CPUs) having one or more processing cores.
[0117] In addition, the simplified computing device 900 shown in FIG. 9 may also include other components such as a communications interface 930. The simplified computing device 900 may also include one or more conventional computer input devices 940 (e.g., pointing devices, keyboards, audio (e.g., voice) input devices, video input devices, haptic input devices, gesture recognition devices, devices for receiving wired or wireless data transmissions, and the like). The simplified computing device 900 may also include other optional components such as one or more conventional computer output devices 950 (e.g., display device(s) 955, audio output devices, video output devices, devices for transmitting wired or wireless data transmissions, and the like). Note that typical communications interfaces 930, input devices 940, output devices 950, and storage devices 960 for general-purpose computers are well known to those skilled in the art, and will not be described in detail herein.
[0118] The simplified computing device 900 shown in FIG. 9 may also include a variety of computer-readable media. Computer-readable media can be any available media that can be accessed by the computer 900 via storage devices 960, and can include both volatile and nonvolatile media that is either removable 970 and/or non-removable 980, for storage of information such as computer-readable or computer-executable instructions, data structures, program modules, or other data. Computer-readable media includes computer storage media and communication media. Computer storage media refers to tangible computer-readable or machine-readable media or storage devices such as digital versatile disks (DVDs), compact discs (CDs), floppy disks, tape drives, hard drives, optical drives, solid state memory devices, random access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, magnetic cassettes, magnetic tapes, magnetic disk storage, or other magnetic storage devices.
[0119] Retention of information such as computer-readable or computer-executable instructions, data structures, program modules, and the like, can also be accomplished by using any of a variety of the aforementioned communication media (as opposed to computer storage media) to encode one or more modulated data signals or carrier waves, or other transport mechanisms or communications protocols, and can include any wired or wireless information delivery mechanism. Note that the terms "modulated data signal" or "carrier wave" generally refer to a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. For example, communication media can include wired media such as a wired network or direct-wired connection carrying one or more modulated data signals, and wireless media such as acoustic, radio frequency (RF), infrared, laser, and other wireless media for transmitting and/or receiving one or more modulated data signals or carrier waves.
[0120] Furthermore, software, programs, and/or computer program products embodying some or all of the various disaster recovery technique embodiments described herein, or portions thereof, may be stored, received, transmitted, or read from any desired combination of computer-readable or machine-readable media or storage devices and communication media in the form of computer-executable instructions or other data structures.
[0121] Finally, the disaster recovery technique embodiments described herein may be further described in the general context of computer-executable instructions, such as program modules, being executed by a computing device. Generally, program modules include routines, programs, objects, components, data structures, and the like, that perform particular tasks or implement particular abstract data types. The disaster recovery technique embodiments may also be practiced in distributed computing environments where tasks are performed by one or more remote processing devices, or within a cloud of one or more devices, that are linked through one or more communications networks. In a distributed computing environment, program modules may be located in both local and remote computer storage media including media storage devices. Additionally, the aforementioned instructions may be implemented, in part or in whole, as hardware logic circuits, which may or may not include a processor.

Claims

1. A computer-implemented process for ensuring the availability of an application being hosted on a current data center, comprising:
using one or more server computers in the current data center to perform the following process actions:
selecting an option for replicating session state data for the application, said selection being made from a set of different session state data replication options each of which comprises different performance and resource cost trade-offs, said selected option determining how said data is to be replicated; and
implementing said selected option, the implementation resulting in said data being replicated outside of the current data center, thus ensuring that said data remains available in the event that the current data center goes offline, while increasing the performance of a given interactive session with the application, or decreasing the resource costs for the interactive session, or both.
2. The process of Claim 1, wherein the current data center is serving an interactive session between the application and a client computer, said selected option comprises storing the most recent session state data for the interactive session on the client computer, and the action of implementing said selected option comprises the actions of:
transmitting a message to the client computer specifying said selected option;
receiving client request messages from the client computer, each of the client request messages comprising the most recent session state data for the interactive session; and
adding the most recent session state data for the interactive session to each application response message that the application transmits to the client computer.
3. The process of Claim 1, wherein the current data center is serving an interactive session between the application and a client computer, said selected option comprises synchronously storing the most recent session state data for the interactive session in one or more backup data centers, and the action of implementing said selected option comprises the actions of:
uploading the most recent session state data for the interactive session to the backup data centers whenever the session state changes; and waiting to respond to a client request message received from the client computer until a confirmation message is received from at least one of the backup data centers indicating that it has successfully stored the most recent session state data.
4. The process of Claim 1, wherein the current data center is serving an interactive session between the application and a client computer, said selected option comprises storing an encoded version of the most recent session state data for the interactive session on the client computer, said encoding reducing the data size, and the action of implementing said selected option comprises the actions of:
transmitting a message to the client computer specifying said selected option;
receiving client request messages from the client computer, each of the client request messages comprising an encoded version of the most recent session state data for the interactive session; and
adding an encoded version of the most recent session state data for the interactive session to each application response message that the application transmits to the client computer.
5. The process of Claim 1, wherein the current data center is serving an interactive session between the application and a client computer, said selected option comprises asynchronously storing the most recent session state data for the interactive session in one or more backup data centers, and the action of implementing said selected option comprises the actions of:
uploading the most recent session state data for the interactive session to the backup data centers whenever the session state changes; and
responding to a client request message received from the client computer before a confirmation message is received from at least one of the backup data centers indicating that it has successfully stored the most recent session state data, the client computer waiting to expose said response to a user until said confirmation message is received from at least one of the backup data centers.
6. The process of Claim 1, wherein the current data center is serving an interactive session between the application and a client computer, and the action of selecting an option for replicating session state data for the application comprises the actions of:
estimating the latency cost to the client computer associated with each of the different session state data replication options; and
selecting a one of said options having the smallest estimated latency cost to the client computer as the option to be used for replicating the session state data for the interactive session.
7. The process of Claim 1, wherein the current data center is serving an interactive session between the application and a client computer, and the action of selecting an option for replicating session state data for the application comprises the actions of:
inputting one or more performance constraints that are specified by the application, said performance constraints comprising a maximum allowable latency cost to the client computer associated with replicating the session state data for the interactive session;
inputting one or more resource cost constraints that are specified by the application;
estimating the latency cost to the client computer associated with each of the different session state data replication options;
identifying ones of said options whose estimated latency cost to the client computer is less than or equal to said maximum allowable latency cost;
selecting a one of said identified options that meets said resource cost constraints as the option to be used for replicating the session state data for the interactive session.
8. The process of Claim 1, wherein the current data center is serving an interactive session between the application and a client computer, and the action of selecting an option for replicating session state data for the application comprises the actions of:
inputting one or more resource cost constraints that are specified by the application;
estimating the latency cost to the client computer associated with each of the different session state data replication options; identifying ones of said options that meet the resource cost constraints; and selecting the one of said identified options having the smallest estimated latency cost to the client computer as the option to be used for replicating the session state data for the interactive session.
9. A computer-implemented process for providing disaster recovery for an application being hosted on a current data center, comprising:
using a client computer to perform the following process actions:
receiving a message from the current data center specifying an option for replicating session state data for the application that has been selected for the client computer, said selection having been made from a set of different session state data replication options each of which comprises different performance and resource cost tradeoffs, said specified option determining how said data is to be replicated; and
implementing said specified option, the implementation resulting in said data being replicated outside of the current data center, thus ensuring that said data remains available to the client computer in the event that the current data center goes offline, while increasing the performance of a given interactive session with the application, or decreasing the resource costs for the interactive session, or both.
10. The process of Claim 9, wherein the current data center is serving an interactive session between the application and the client computer, said specified option comprises storing the most recent session state data for the interactive session on the client computer, and the action of implementing said specified option comprises the actions of:
receiving application response messages from the application, each of the application response messages comprising the most recent session state data for the interactive session;
extracting the most recent session state data for the interactive session from each of the application response messages;
storing said extracted data on the client computer;
detecting when the current data center has gone offline and thus has become unable to continue serving the interactive session; and
whenever it is detected that the current data center has gone offline, transmitting a session resume request message to a backup data center that has been selected to take over serving the interactive session, said message comprising said stored data.
PCT/US2014/063614 2013-11-11 2014-11-03 Geo-distributed disaster recovery for interactive cloud applications WO2015069584A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP14805405.9A EP3069496B1 (en) 2013-11-11 2014-11-03 Geo-distributed disaster recovery for interactive cloud applications
CN201480061769.2A CN105765949B (en) 2013-11-11 2014-11-03 The method of geographically distributed disaster recovery for interactive cloud application
BR112016010037A BR112016010037A8 (en) 2013-11-11 2014-11-03 process for ensuring the availability of an application being hosted in a current data center, process for providing disaster recovery for an application being hosted in a current data center and process for supplying and using spare data center capacity over a network of data centers

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US14/076,715 2013-11-11
US14/076,715 US10594784B2 (en) 2013-11-11 2013-11-11 Geo-distributed disaster recovery for interactive cloud applications

Publications (1)

Publication Number Publication Date
WO2015069584A1 true WO2015069584A1 (en) 2015-05-14

Family

ID=51999514

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2014/063614 WO2015069584A1 (en) 2013-11-11 2014-11-03 Geo-distributed disaster recovery for interactive cloud applications

Country Status (5)

Country Link
US (1) US10594784B2 (en)
EP (1) EP3069496B1 (en)
CN (1) CN105765949B (en)
BR (1) BR112016010037A8 (en)
WO (1) WO2015069584A1 (en)

Families Citing this family (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
FR3014623B1 (en) * 2013-12-06 2016-01-22 Airbus Operations Sas AIRCRAFT COMMUNICATION SYSTEM HAVING A REDUNDANT NETWORK
US9785505B1 (en) * 2014-05-12 2017-10-10 Amazon Technologies, Inc. Communicative building management system for data loss prevention
US10383059B2 (en) * 2014-05-23 2019-08-13 General Motors Llc Vehicle telematics unit power management
US20150348177A1 (en) * 2014-05-29 2015-12-03 International Business Machines Corporation Managing lease transactions in distributed systems
US10650085B2 (en) 2015-03-26 2020-05-12 Microsoft Technology Licensing, Llc Providing interactive preview of content within communication
JP6540356B2 (en) * 2015-08-10 2019-07-10 富士通株式会社 System replication control device and system replication control method
US9971664B2 (en) * 2015-08-27 2018-05-15 Vmware, Inc. Disaster recovery protection based on resource consumption patterns
US9875373B2 (en) 2015-09-28 2018-01-23 International Business Machines Corporation Prioritization of users during disaster recovery
US10962780B2 (en) * 2015-10-26 2021-03-30 Microsoft Technology Licensing, Llc Remote rendering for virtual images
US9430337B1 (en) 2016-01-07 2016-08-30 International Business Machines Corporation Disaster recovery as a dynamic service
US10324635B1 (en) * 2016-03-22 2019-06-18 EMC IP Holding Company LLC Adaptive compression for data replication in a storage system
US10565058B1 (en) 2016-03-30 2020-02-18 EMC IP Holding Company LLC Adaptive hash-based data replication in a storage system
US10567460B2 (en) * 2016-06-09 2020-02-18 Apple Inc. Managing data using a time-based directory structure
EP3491801B1 (en) * 2016-07-28 2021-01-13 Koninklijke Philips N.V. Identifying a network node to which data will be replicated
US10116569B1 (en) 2016-08-19 2018-10-30 EMC IP Holding Company LLC Embedded diagnostic tool for prediction of problems relating to network bandwidth
US10614043B2 (en) * 2016-09-30 2020-04-07 Adobe Inc. Document replication based on distributional semantics
US10241848B2 (en) 2016-09-30 2019-03-26 Microsoft Technology Licensing, Llc Personalized diagnostics, troubleshooting, recovery, and notification based on application state
US10476768B2 (en) * 2016-10-03 2019-11-12 Microsoft Technology Licensing, Llc Diagnostic and recovery signals for disconnected applications in hosted service environment
CN107959693B (en) * 2016-10-14 2022-04-15 中兴通讯股份有限公司 Calling method, synchronous server and calling system
US10447672B2 (en) * 2016-11-01 2019-10-15 Salesforce.Com, Inc. Facilitating encrypted persistent storage in browsers
US11080632B2 (en) * 2016-11-29 2021-08-03 International Business Machines Corporation Optimization of steady state cost for multi-site high availability application deployment of management and managed intrastructure
CN108632067B (en) * 2017-03-21 2020-12-08 华为技术有限公司 Disaster recovery deployment method, device and system
CN108287762B (en) * 2018-01-09 2022-04-12 福建星瑞格软件有限公司 Distributed computing interactive mode use resource optimization method and computer equipment
US10452296B1 (en) 2018-03-23 2019-10-22 Amazon Technologies, Inc. Accelerated volumes
US11023157B2 (en) 2018-04-30 2021-06-01 Amazon Technologies, Inc. Intermediary duplication to facilitate copy requests in distributed storage systems
US10459655B1 (en) 2018-04-30 2019-10-29 Amazon Technologies, Inc. Rapid volume backup generation from distributed replica
US11343314B1 (en) 2018-04-30 2022-05-24 Amazon Technologies, Inc. Stream-based logging for distributed storage systems
CN108833479B (en) * 2018-05-18 2021-02-09 吉林亿联银行股份有限公司 Data synchronization method and device
US10931750B1 (en) * 2018-07-30 2021-02-23 Amazon Technologies, Inc. Selection from dedicated source volume pool for accelerated creation of block data volumes
US10956442B1 (en) 2018-07-30 2021-03-23 Amazon Technologies, Inc. Dedicated source volume pool for accelerated creation of block data volumes from object data snapshots
US11068192B1 (en) 2019-03-26 2021-07-20 Amazon Technologies, Inc. Utilizing mutiple snapshot sources for creating new copy of volume in a networked environment wherein additional snapshot sources are reserved with lower performance levels than a primary snapshot source
CN109981777A (en) * 2019-03-27 2019-07-05 深圳市网心科技有限公司 Copy maintaining method, server, copy memory node and system based on CDN
US10983719B1 (en) 2019-03-28 2021-04-20 Amazon Technologies, Inc. Replica pools to support volume replication in distributed storage systems
CN110009244B (en) * 2019-04-12 2021-04-20 西安交通大学 Regional comprehensive energy system optimization scheduling method considering disaster recovery
US11269689B2 (en) * 2019-08-02 2022-03-08 International Business Machines Corporation Distribution of components of displayed data between a server and a client based on server and client load factors
US11544149B2 (en) * 2020-01-07 2023-01-03 Oracle International Corporation System and method for improved fault tolerance in a network cloud environment
US11323507B2 (en) * 2020-04-07 2022-05-03 Supercell Oy Server system and method of managing server system
US11899625B2 (en) * 2021-06-30 2024-02-13 EMC IP Holding Company LLC Systems and methods for replication time estimation in a data deduplication system
US11886307B2 (en) * 2021-07-19 2024-01-30 Vmware, Inc. Site locality support for file services in a stretched cluster environment
US11853559B1 (en) 2022-09-22 2023-12-26 International Business Machines Corporation Mirror write consistency check policy for logical volume manager systems

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060168334A1 (en) * 2005-01-25 2006-07-27 Sunil Potti Application layer message-based server failover management by a network element
US20130054822A1 (en) * 2011-08-30 2013-02-28 Rajiv P. Mordani Failover Data Replication with Colocation of Session State Data

Family Cites Families (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6076108A (en) * 1998-03-06 2000-06-13 I2 Technologies, Inc. System and method for maintaining a state for a user session using a web system having a global session server
AU2000266395A1 (en) * 2000-08-11 2002-02-25 George Mount Server-side session management
US20030110266A1 (en) * 2001-12-10 2003-06-12 Cysive, Inc. Apparatus and method of using session state data across sessions
US7877435B2 (en) 2002-06-20 2011-01-25 International Business Machines Corporation Method and system for transaction pipeline decomposition
US20050010588A1 (en) * 2003-07-08 2005-01-13 Zalewski Stephen H. Method and apparatus for determining replication schema against logical data disruptions
US8850141B2 (en) * 2003-07-15 2014-09-30 International Business Machines Corporation System and method for mirroring data
US8359491B1 (en) * 2004-03-30 2013-01-22 Symantec Operating Corporation Disaster recovery rehearsal using copy on write
US7747760B2 (en) * 2004-07-29 2010-06-29 International Business Machines Corporation Near real-time data center switching for client requests
US8281014B2 (en) * 2004-12-28 2012-10-02 Sap Ag Session lifecycle management within a multi-tiered enterprise network
US8064356B1 (en) * 2005-01-11 2011-11-22 Verizon Services Corp. System and methods for measuring network performance
US7761435B2 (en) * 2005-04-29 2010-07-20 Sap Ag External persistence of session state information
US8024566B2 (en) * 2005-04-29 2011-09-20 Sap Ag Persistent storage implementations for session data within a multi-tiered enterprise network
US8199654B2 (en) * 2005-06-21 2012-06-12 Alcatel Lucent Method and apparatus for providing end-to-end high quality services based on performance characterizations of network conditions
US7651593B2 (en) 2005-12-19 2010-01-26 Commvault Systems, Inc. Systems and methods for performing data replication
US8051179B2 (en) * 2006-02-01 2011-11-01 Oracle America, Inc. Distributed session failover
US8024439B2 (en) * 2006-03-17 2011-09-20 Microsoft Corporation Server session management application program interface and schema
US7711848B2 (en) * 2006-06-15 2010-05-04 Oracle International Corporation System using session initiation protocol for seamless network switching in a media streaming session
US7751320B2 (en) * 2006-09-30 2010-07-06 Alcatel-Lucent Usa Inc. Admission control of sessions with preference lists
US8725890B2 (en) * 2006-10-31 2014-05-13 Thomson Licensing Data recovery in heterogeneous networks using peer's cooperative networking
US7685227B2 (en) 2006-11-10 2010-03-23 Gerber Robert H Message forwarding backup manager in a distributed server system
US7844851B2 (en) 2006-12-13 2010-11-30 Oracle International Corporation System and method for protecting against failure through geo-redundancy in a SIP server
JP4982304B2 (en) * 2007-09-04 2012-07-25 株式会社日立製作所 Storage system that understands the occurrence of power failure
US7870095B2 (en) * 2007-12-03 2011-01-11 International Business Machines Corporation Apparatus, system, and method for replication of data management information
WO2010102084A2 (en) 2009-03-05 2010-09-10 Coach Wei System and method for performance acceleration, data protection, disaster recovery and on-demand scaling of computer applications
US20100228824A1 (en) * 2009-03-06 2010-09-09 Cisco Technology, Inc. Distributed server selection for online collaborative computing sessions
US9191322B2 (en) * 2009-05-19 2015-11-17 Beaumaris Networks Inc. Methods, apparatus and computer readable medium for managed adaptive bit rate for bandwidth reclamation
US9537957B2 (en) * 2009-09-02 2017-01-03 Lenovo (Singapore) Pte. Ltd. Seamless application session reconstruction between devices
US8694469B2 (en) * 2009-12-28 2014-04-08 Riverbed Technology, Inc. Cloud synthetic backups
US8589565B2 (en) 2010-05-24 2013-11-19 Microsoft Corporation Client-server session parallelism
US9686355B2 (en) * 2010-12-20 2017-06-20 Microsoft Technology Licensing, Llc Third party initiation of communications between remote parties
US8539338B2 (en) * 2011-04-11 2013-09-17 Microsoft Corporation Cooperative rendering cache for mobile browser
US8954786B2 (en) * 2011-07-28 2015-02-10 Oracle International Corporation Failover data replication to a preferred list of instances
US20130073670A1 (en) 2011-09-15 2013-03-21 Microsoft Corporation Geo-Migration Of User State
US10152398B2 (en) * 2012-08-02 2018-12-11 At&T Intellectual Property I, L.P. Pipelined data replication for disaster recovery
US9378060B2 (en) * 2012-08-28 2016-06-28 Oracle International Corporation Runtime co-location of executing logic and frequently-accessed application data
US8874506B2 (en) * 2012-09-10 2014-10-28 Oracle International Corporation Preventing database replication conflicts in a distributed environment
US8977598B2 (en) * 2012-12-21 2015-03-10 Zetta Inc. Systems and methods for on-line backup and disaster recovery with local copy
US9075704B2 (en) * 2013-01-25 2015-07-07 Hewlett-Packard Development Company, L.P. Mitigating risks during a high availibility and disaster recovery (HA/DR) rehearsal
US9465855B2 (en) * 2013-10-22 2016-10-11 International Business Machines Corporation Maintaining two-site configuration for workload availability between sites at unlimited distances for products and services

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060168334A1 (en) * 2005-01-25 2006-07-27 Sunil Potti Application layer message-based server failover management by a network element
US20130054822A1 (en) * 2011-08-30 2013-02-28 Rajiv P. Mordani Failover Data Replication with Colocation of Session State Data

Also Published As

Publication number Publication date
EP3069496A1 (en) 2016-09-21
CN105765949B (en) 2019-11-05
CN105765949A (en) 2016-07-13
US10594784B2 (en) 2020-03-17
EP3069496B1 (en) 2020-02-12
BR112016010037A8 (en) 2020-04-14
US20150134723A1 (en) 2015-05-14

Similar Documents

Publication Publication Date Title
EP3069496B1 (en) Geo-distributed disaster recovery for interactive cloud applications
US10887416B2 (en) Efficient high availability and storage efficiency in a multi-site object storage environment
US10489422B2 (en) Reducing data volume durability state for block-based storage
US9983825B2 (en) Efficient data volume replication for block-based storage
US10715460B2 (en) Opportunistic resource migration to optimize resource placement
Abolfazli et al. Cloud-based augmentation for mobile devices: motivation, taxonomies, and open challenges
US10482104B2 (en) Zero-data loss recovery for active-active sites configurations
US9350682B1 (en) Compute instance migrations across availability zones of a provider network
US10924542B2 (en) Content delivery system
US9535793B2 (en) Method and system for data migration
RU2632125C1 (en) Method and system for tasks processing in cloud service
US20140188801A1 (en) Method and system for intelligent load balancing
US10505862B1 (en) Optimizing for infrastructure diversity constraints in resource placement
US10185596B2 (en) Cloud book registry for cloud service providers wherein the consumer can access the profile for each cloud service provider and service usage of other consumers
US10956880B2 (en) Mixed deployment architecture for distributed services
US10469571B2 (en) Block allocation based on server utilization
US8984030B2 (en) Journaling and integrity in mobile clouded collaborative spaces
FR3099613A1 (en) AUTOMATIC LEARNING SYSTEMS AND METHODS FOR PLACING DATA IN DISTRIBUTED STORAGE
US11061727B1 (en) Preventing performance degradation over time in a cache
US10789273B1 (en) Resource identifier allocation during network partitions
US10959059B1 (en) Providing intelligent content recommendations to groups of people
US20160323390A1 (en) Condition-based online communication collaboration
US11102256B2 (en) System and method for dynamic cloud formations for stream computing
US20190141112A1 (en) Cognitive storage management for a mobile computing device
US11416635B1 (en) Pseudonymous browsing mode

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14805405

Country of ref document: EP

Kind code of ref document: A1

DPE2 Request for preliminary examination filed before expiration of 19th month from priority date (pct application filed from 20040101)
REEP Request for entry into the european phase

Ref document number: 2014805405

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2014805405

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE

REG Reference to national code

Ref country code: BR

Ref legal event code: B01A

Ref document number: 112016010037

Country of ref document: BR

ENP Entry into the national phase

Ref document number: 112016010037

Country of ref document: BR

Kind code of ref document: A2

Effective date: 20160504