WO2013163221A1 - Systems and methods for adaptive streaming with augmented video stream transitions - Google Patents

Systems and methods for adaptive streaming with augmented video stream transitions Download PDF

Info

Publication number
WO2013163221A1
WO2013163221A1 PCT/US2013/037851 US2013037851W WO2013163221A1 WO 2013163221 A1 WO2013163221 A1 WO 2013163221A1 US 2013037851 W US2013037851 W US 2013037851W WO 2013163221 A1 WO2013163221 A1 WO 2013163221A1
Authority
WO
WIPO (PCT)
Prior art keywords
stream
image
video data
pictures
group
Prior art date
Application number
PCT/US2013/037851
Other languages
French (fr)
Inventor
Christopher Dow
Alexander Ivanov
Kourosh Soroushian
Original Assignee
Divx, Llc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/457,911 external-priority patent/US9060184B2/en
Priority claimed from US13/457,901 external-priority patent/US20130287092A1/en
Application filed by Divx, Llc filed Critical Divx, Llc
Publication of WO2013163221A1 publication Critical patent/WO2013163221A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44209Monitoring of downstream path of the transmission network originating from a server, e.g. bandwidth variations of a wireless network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/426Internal components of the client ; Characteristics thereof
    • H04N21/42653Internal components of the client ; Characteristics thereof for processing graphics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/63Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
    • H04N21/637Control signals issued by the client directed to the server or network components
    • H04N21/6373Control signals issued by the client directed to the server or network components for rate control, e.g. request to the server to modify its transmission rate
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/85406Content authoring involving a specific file format, e.g. MP4 format

Abstract

Systems and methods for adaptive streaming with augmented video stream transitions in accordance with embodiments of the invention are disclosed. In one embodiment of the invention, a network client includes memory containing a video decoder application, an image processing application, and a client application and a processor, wherein the video decoder application configures the processor to decode video data for display, wherein the image processor application configures the processor to render an image for display, wherein the client application configures the processor to receive a first stream of video data, play the first stream of video data, detect a change in streaming conditions, request a second stream of video data, render an image for display, receive the second stream of video data, and play the second stream of video data, and coordinates the display of the image between the first stream of video and the second video stream.

Description

Systems and Methods for Adaptive Streaming with
Augmented Video Stream Transitions
FIELD OF THE INVENTION
[0001] The present invention is directed, in general, to systems and methods for streaming data over a network and more specifically to systems and methods for streaming video data over a network utilizing adaptive streaming techniques.
BACKGROUND
[0002] The term streaming media describes the playback of media on a playback device, where the media is stored on a server and continuously sent to the playback device over a network during playback. Typically, the playback device stores a sufficient quantity of media in a buffer at any given time during playback to prevent disruption of playback due to the playback device completing playback of all the buffered media prior to receipt of the next portion of media. Adaptive bit rate streaming or adaptive streaming involves detecting the present streaming conditions (e.g. the user's network bandwidth and CPU capacity) in real time and adjusting the quality of the streamed media accordingly. Streaming video over the Internet has become a phenomenon in modern times. Many popular websites, such as YouTube, a service of Google, Inc. of Mountain View, California, and WatchESPN, a service of ESPN of Bristol, Connecticut, utilize streaming video in order to provide video and television programming to consumers via the Internet.
[0003] Scalable Video Coding (SVC) is an extension of the H.264/MPEG-4 AVC video compression standard, which is specified by the ITU-T H.264 standard by the International Telecommunication Union Telecommunication Standardization Sector of Geneva, Switzerland. SVC enables the encoding of a video bitstream that additionally contains one or more sub-bitstreams. The sub-bitstreams are derived from the video bitstream by dropping packets of data from the video bitstream, resulting in a sub- bitstream of lower quality and lower bandwidth than the original video bitstream. SVC supports three forms of scaling a video bitstream into sub-bitstreams: temporal scaling, spatial scaling, and quality scaling. Each of these scaling techniques can be used individually or combined depending on the specific video system. [0004] In adaptive streaming systems, the source media is typically stored on a media server as a top level index file pointing to a number of alternate streams that contain the actual video and audio data. Each stream is typically stored in one or more container files. Different adaptive streaming solutions typically utilize different index and media containers. The Matroska container is a media container developed as an open standard project by the Matroska non-profit organization of Aussonne, France. The Matroska container is based upon Extensible Binary Meta Language (EBML), which is a binary derivative of the Extensible Markup Language (XML). Decoding of the Matroska container is supported by many consumer electronics (CE) devices. The DivX Plus file format developed by DivX, LLC of San Diego, California utilizes an extension of the Matroska container format, including elements that are not specified within the Matroska format.
[0005] A digital image is a digital representation of an image. Digital images usually include a fixed number of pixels arranged in rows and columns in order to form a two- dimensional image. Digital images may exist as uncompressed bitmaps, or may be compressed using a variety of techniques and stored in a variety of formats, such as the Graphics Interchange Format (GIF), developed by CompuServe of Columbus, Ohio, the Joint Photographic Experts Group format (JPEG), developed by the Joint Photographic Experts Group of the International Organization for Standardization (ISO) and the International Electrotechnical Commission (IEC), both of Geneva, Switzerland, or the Portable Network Graphics (PNG) format, standardized as ISO/IEC 15948 by the ISO and IEC of Geneva, Switzerland. Other image compression techniques can be utilized as appropriate to the requirements of a specific application.
SUMMARY OF THE INVENTION
[0006] Systems and methods for adaptive streaming with augmented video stream transitions in accordance with embodiments of the invention are disclosed. In one embodiment of the invention, a network client includes memory containing a video decoder application, an image processing application, and a client application and a processor, wherein the video decoder application configures the processor to decode video data for display, wherein the image processor application configures the processor to render an image for display using image data, wherein the client application configures the processor to receive a first stream of video data, play the first stream of video data using the video decoder, detect a change in streaming conditions, request a second stream of video data, render an image for display using the image processor, receive the second stream of video data, and play the second stream of video data using the video decoder, and wherein the client application coordinates the display of the rendered image between a last frame of the first stream of video played back by the network client and a first frame of the second stream of video data played back by the network client.
[0007] In another embodiment of the invention, the client application configures the processor to receive the image for display from a media server.
[0008] In an additional embodiment of the invention the first stream of video data includes a first group of pictures, where the first group of pictures includes a first frame and a last frame and the client application configures the processor to generate the image for display using the first group of pictures.
[0009] In yet another additional embodiment of the invention, the image for display is generated using the last frame of the first group of pictures.
[0010] In still another additional embodiment of the invention, the second stream of video data includes a second group of pictures, where the second group of pictures includes a first frame and a last frame and the client application configures the processor to generate the image for display using the second group of pictures.
[0011] In yet still another additional embodiment of the invention, the image for display is generated using the first frame of the second group of pictures.
[0012] In yet another embodiment of the invention, the image for display is pre- computed.
[0013] In still another embodiment of the invention, the image for display is determined in real time.
[0014] In yet still another embodiment of the invention, the video decoder is implemented using the processor.
[0015] In yet another additional embodiment of the invention, the image processor is implemented using the processor. [0016] Still another embodiment of the invention includes augmented video transitions using a network client configured to receive streams of video data, including receiving a first stream of video data using a network client, playing the first stream of video data using the network client, detecting a change in streaming conditions using the network client, requesting a second stream of video data using the network client, rendering an image for display using the network client, and receiving a second stream of video data using the network client, and playing the second stream of video data using the network client, wherein the displayed image is displayed between a last frame of the first stream of video data displayed and a first frame of the second video stream displayed.
[0017] In yet another additional embodiment of the invention, augmented video transitions further include receiving the image for display using the network client.
[0018] In still another additional embodiment of the invention, the first stream of video data includes a first group of pictures, where the first group of pictures includes a first frame and a last frame and generating the image for display further includes using the first group of pictures to generate the image for display.
[0019] In yet still another additional embodiment of the invention, generating the image for display utilizes the last frame of the first group of pictures.
[0020] In yet another embodiment of the invention, the second stream of video data includes a second group of pictures, where the second group of pictures includes a first frame and a last frame and generating the image for display further includes using the second group of pictures to generate the image for display.
[0021] In still another embodiment of the invention, generating the image for display utilizes the first frame of the second group of pictures.
[0022] In yet still another embodiment of the invention, the image for display is pre-computed.
[0023] In yet another additional embodiment of the invention, augmented video transitions further include determining the image for display in real time using the network client.
[0024] In still another additional embodiment of the invention, augmented video transitions further include resizing the image for display using the network client. [0025] Still another embodiment of the invention includes a machine readable medium containing processor instructions, where execution of the instructions by a processor causes the processor to perform a process including receiving a first stream of video data, playing the first stream of video data, detecting a change in streaming conditions, requesting a second stream of video data, rendering an image for display, receiving a second stream of video data, and playing the second stream of video data, wherein the displayed image is displayed between a last frame of the first stream of video data displayed and a first frame of the second video stream displayed.
BRIEF DESCRIPTION OF THE DRAWINGS
[0026] FIG. 1 is a system diagram of a system for streaming video data in accordance with an embodiment of the invention.
[0027] FIG. 2 conceptually illustrates a network client configured to perform dynamic video scaling in accordance with an embodiment of the invention.
[0028] FIG. 3 is a flow chart illustrating a process for server-based augmented video stream transitions in accordance with an embodiment of the invention.
[0029] FIG. 4 is a flow chart illustrating a process for client-based augmented video stream transitions in accordance with an embodiment of the invention.
DETAILED DESCRIPTION
[0030] Turning now to the drawings, systems and methods for adaptive streaming with augmented video stream transitions in accordance with embodiments of the invention are disclosed. Adaptive streaming systems are configured to stream multimedia content encoded at different maximum bitrates and resolutions over a network, such as the Internet. Adaptive streaming systems stream the highest quality multimedia content which can be supported based upon current streaming conditions. Multimedia content typically includes video and audio data, subtitles, and other related metadata. In order to provide the highest quality video experience independent of the network data rate, adaptive bitrate streaming systems are configured to switch between the available sources of video data throughout the delivery of the video data according to a variety of factors, including, but not limited to, the available network data rate and video decoder performance. When streaming conditions deteriorate, an adaptive bitrate streaming system typically attempts to switch to multimedia stream(s) encoded at lower maximum bitrates. In the event that the available network data rate cannot support streaming of the stream(s) encoded at the lowest maximum bitrate, then playback is often disrupted until a sufficient amount of content can be buffered to restart playback. When displaying a video stream, adaptive bitrate streaming systems often initialize a video decoder in order to decode and display the video stream. The process of initializing a video decoder may result in unwanted visual disturbances (particularly on resource constrained devices). Adaptive bitrate streaming systems in accordance with embodiments of the invention can utilize augmented video stream transitions to display an image when switching between video streams in order to obscure any visual disruptions and provide an improved user experience. Systems and methods for switching between video streams during playback that can be adapted to include augmented video stream transitions in accordance with embodiments of the invention are described in U.S. Patent Application Serial No. 13/221 ,682 entitled "Systems and Methods for Adaptive Bitrate Streaming of Media Stored in Matroska Container Files Using Hypertext Transfer Protocol" to Braness et al., filed August 30, 201 1 , the disclosure of which is incorporated by reference herein in its entirety. A variety of methods may be utilized in accordance with embodiments of the invention in order to generate the image displayed during video stream transitions, such as retrieving the image from a media server or capturing the currently displayed image on a network client. Systems and methods for adaptive bitrate streaming with augmented video stream transitions in accordance with embodiments of the invention are discussed further below.
System Overview
[0031] Video data networks in accordance with embodiments of the invention are configured to adapt the bitrate of the video transmitted to network clients based on streaming conditions. A video data network in accordance with an embodiment of the invention is illustrated in FIG. 1 . The illustrated video data network 10 includes a video source 100. In a number of embodiments of the invention, the video source 100 contains pre-encoded video data encoded at multiple bitrates and resolutions for adaptive bitrate streaming. In several embodiments of the invention, the video source encodes video data in real time. In many embodiments of the invention, the video source contains video data encoded utilizing Scalable Video Coding (SVC). In a number of embodiments, the video source contains multiple streams with equal timelines as video data. The video source 100 is connected to a network renderer 102. In many embodiments, the video source 100 and the network renderer 102 are implemented using a media server. The network renderer 102 is connected to a plurality of network clients 104 via a network 1 10. The network renderer 102 is configured to stream video data to one or more network clients 104.
[0032] In many embodiments of the invention, the network renderer 102 is implemented using a single machine. In several embodiments of the invention, the network renderer is implemented using a plurality of machines. In many embodiments of the invention, the network renderer and the video source are implemented using a media server. In many embodiments, the network 1 10 is the Internet. In several embodiments, the network 1 10 is any IP network. In a number of embodiments, the network 1 10 is a cellular data network.
[0033] The network clients 104 each contain a video decoder 106, an image processor 107, and a client application 108. Network clients 104 are configured to decode video streams using the video decoder 106 and to display images using the image processor 107. In many embodiments, the network clients 104 are configured to switch between video streams based on a variety of factors, including the performance of the network connection and the performance of the video decoder 106. In several embodiments, switching between video streams causes visual disturbances. In a number of embodiments, the network clients 104 are configured to retrieve an image from the network renderer 102 and display the retrieved image using the image processor 107 while switching between video streams. In several embodiments, the network clients 104 are configured to retrieve an image from the video decoder 106 and display the retrieved image using the image processor 107 while switching between video streams. In many embodiments, the client application 108 is configured to control the creation and termination of the video decoder 106 and/or the image processor 107. In several embodiments, the client application controls the playback of video streams using the video decoder 106 and/or the image processor 107. In a number of embodiments, the client application 108 controls the display of an image between the last frame of a first stream of video played back by the network client and the first frame of a second video stream played back by the network client.
[0034] In many embodiments of the invention, network clients can include consumer electronics devices such as DVD players, Blu-ray players, televisions, set top boxes, video game consoles, tablets, and other devices that are capable of connecting to a server and playing back encoded media. The basic architecture of a network client in accordance with an embodiment of the invention is illustrated in FIG. 2. The network client 200 includes a processor 210 in communication with non-volatile memory 230 and volatile memory 220. In the illustrated embodiment, the non-volatile memory includes a video decoder 232 that configures the processor to decode scalable video data, an image processor 234 which is configured to display images, and a client application 236 configured to create, operate, and terminate the video decoder 232 and/or the image processor 234. In many embodiments of the invention, the image processor 234 is configured to decoded encoded images for display. In several embodiments, the video decoder 232, the image processor 234, and/or the client application 236 are implemented using the processor 210. The network client 200 also includes a network interface 240 configured to send and receive data over a network connection. In a number of embodiments, the network interface 240 is in communication with the processor 210, the non-volatile memory 230, and the volatile memory 220.
[0035] In many embodiments of the invention, the network client 200 includes multiple processors. In several embodiments, the video decoder 232, the image processor 234, and/or the client application 236 are utilize separate processors. In a number of embodiments, the video decoder 232, the image processor 234, and/or the client application 236 are implemented using dedicated hardware. Although a specific network client architecture is illustrated in FIG. 2, any of a variety of architectures including architectures where the video decoder is located on disk or some other form of storage and is loaded into volatile memory at runtime can be utilized to implement network clients for use in adaptive video data streaming systems with augmented video transitions in accordance with embodiments of the invention.
[0036] Although a specific architecture of a video data network is shown in FIG. 1 , other implementations appropriate to a specific application can be utilized in accordance with embodiments of the invention. Processes for streaming video data with augmented video stream transitions in accordance with embodiments of the invention are discussed further below.
Server-Based Augmented Video Stream Transitions
[0037] While playing back streamed video data, network clients often switch between video streams based on network conditions or video decoder performance in order to ensure smooth video playback. However, many network clients experience delays or visual disturbances when transitioning between video streams, such as when the video decoder is initialized in order to display the new video stream or when the network data rate drops to a point where the throughput of the network is insufficient to support streaming video data and the system buffers data sufficient to resume playback of the video stream. A process for sever-based augmented video stream transitions for use in adaptive streaming systems with augmented video stream transitions in accordance with an embodiment of the invention is illustrated in FIG. 3. The process 300 includes streaming (310) video. In many embodiments, the video is streamed (310) between a media server and a network client. The video is displayed (31 1 ). When a determination is made that a video stream change is needed (312), the process initiates a stream transition. In a number of embodiments, the video stream transition is based on network conditions and/or the performance of the network client. The network client retrieves (314) an image from the media server. In several embodiments, the image retrieved (314) corresponds to the first frame of a group of pictures. In many embodiments, the image retrieved (314) corresponds to the last frame of a group of pictures. In several embodiments, the image retrieved (314) corresponds to the last frame of video displayed (31 1 ) to the network client. In a number of embodiments, the image retrieved (314) is created by the media server at the time the image is retrieved. In many embodiments, the image retrieved (314) is pre-computed and stored on the media server. In a number of embodiments, the image retrieved (314) is retrieved from a separate image server. In several embodiments, the image retrieved (314) has been created according to the resolution of the streaming video (310). In a number of embodiments, the image retrieved (314) has been created according to the highest resolution of the adaptive stream videos.
[0038] In several embodiments, the network client resizes (315) the image. In many embodiments, the image is resized (315) such that it matches the resolution of the last video frame. In a number of embodiments, the image is resized (315) such that it matches the resolution of the first video frame of the next group of frames. The network client displays (316) the image. In a number of embodiments, the network client displays (316) the image using an image processor. As the image is displayed, the network client (318) prepares to receive media from a second video stream. In a number of embodiments, the second video stream is at a different resolution than the first video stream. In several embodiments, changing (318) the video stream involves initializing a video decoder. Initializing a video decoder can create visual disturbances, such as displaying video artifacts or resetting a display device. In several embodiments, visual disturbances are avoided by displaying an image using an image processor. The network client resumes (320) streaming video and stops displaying the image. In a number of embodiments, by displaying an image while changing to the second video stream, any visual artifacts caused by changing video streams are not displayed on the network client.
[0039] Although a specific method for server-based augmented video stream transitions for use in adaptive streaming systems with augmented video stream transitions is illustrated in FIG. 3, other implementations can be utilized in accordance with embodiments of the invention. Processes for client-based augmented video stream transitions for use in adaptive streaming systems with augmented video stream transitions in accordance with embodiments of the invention are discussed further below. Client-Based Augmented Video Stream Transitions
[0040] While playing back streamed video data, network clients often switch between video streams based on network conditions or video decoder performance in order to ensure smooth video playback. While switching between video streams, augmented video stream transitions may be implemented using the network client. A process for client-based augmented video stream transitions for use in adaptive streaming systems in accordance with an embodiment of the invention is illustrated in FIG. 4. The process 400 includes streaming (410) video. In many embodiments, the video is streamed (410) between a media server and a network client. The video is displayed (41 1 ). When a determination is made that a video stream change is needed (412), the processes initiates a stream transition. In a number of embodiments, the video stream change is based on network conditions and/or the performance of the network client. In several embodiments, a network client may be unable to retrieve an image from a media server for a variety of reasons, including network conditions and/or the performance of the network client. Therefore, the network client captures (414) an image. In several embodiments, the network client always captures an image irrespective of whether the network conditions would support downloading of a pre-generated image from a remote server. In many embodiments, the captured (414) image is the image currently being displayed (41 1 ) by the network client. In several embodiments, the image is captured (414) directly from a display buffer. In many embodiments, the captured (414) image is uncompressed. In several embodiments, the captured (414) image is compressed using an image compression technique, such as (but not limited to) the Graphics Interchange Format (GIF), the Joint Photographic Experts Group (JPEG) format, or the Portable Network Graphics (PNG) format. In a number of embodiments, the image (414) corresponds to a frame in the received streamed video data. In many embodiments, the captured (414) image is the first frame of a group of pictures. In several embodiments, the captured (414) image is the last frame of a group of pictures.
[0041] The network client displays (416) the image. In a number of embodiments, the network client displays (416) the image using an image processor. In many embodiments, an image processor is implemented using a processor configured to display images in a specified format. The network client changes (418) to a second video stream. In a number of embodiments, the second video stream is at a different resolution than the first video stream. In several embodiments, changing (418) the video stream involves initializing a video decoder. In many embodiments, initializing a video decoder creates visual disturbances, such as displaying video artifacts or resetting a display device. In several embodiments, visual disturbances caused by the video decoder do not affect the display of an image using an image processor. The network client resumes (420) streaming video. In a number of embodiments, by displaying an image while changing to the second video stream, any visual artifacts that would otherwise be displayed when changing between video streams are not displayed on the network client. A specific method for client-based augmented video stream transitions for use in adaptive streaming systems with augmented video stream transitions is illustrated in FIG. 4; however, other implementations can be utilized in accordance with embodiments of the invention.
[0042] Although the present invention has been described in certain specific aspects, many additional modifications and variations would be apparent to those skilled in the art. It is therefore to be understood that the present invention may be practiced otherwise than specifically described without departing from the scope and spirit of the present invention. Thus, embodiments of the present invention should be considered in all respects as illustrative and not restrictive. Accordingly, the scope of the invention should be determined not by the embodiments illustrated, but by the appended claims and their equivalents.

Claims

WHAT IS CLAIMED:
1 . A network client, comprising:
memory containing a video decoder application, an image processing application, and a client application; and
a processor;
wherein the video decoder application configures the processor to decode video data for display;
wherein the image processor application configures the processor to render an image for display using image data;
wherein the client application configures the processor to:
receive a first stream of video data;
play the first stream of video data using the video decoder;
detect a change in streaming conditions;
request a second stream of video data;
render an image for display using the image processor;
receive the second stream of video data; and
play the second stream of video data using the video decoder; and wherein the client application coordinates the display of the rendered image between a last frame of the first stream of video data played back by the network client and a first frame of the second stream of video data played back by the network client.
2. The network client of claim 1 , wherein the client application configures the processor to receive the image for display from a media server.
3. The network client of claim 1 , wherein:
the first stream of video data comprises a first group of pictures, where the first group of pictures comprises a first frame and a last frame; and
the client application configures the processor to generate the image for display using the first group of pictures.
4. The network client of claim 3, wherein the image for display is generated using the last frame of the first group of pictures.
5. The network client of claim 1 , wherein:
the second stream of video data comprises a second group of pictures, where the second group of pictures comprises a first frame and a last frame; and
the client application configures the processor to generate the image for display using the second group of pictures.
6. The network client of claim 5, wherein the image for display is generated using the first frame of the second group of pictures.
7. The network client of claim 1 , wherein the image for display is pre- computed.
8. The network client of claim 1 , wherein the image for display is determined in real time.
9. The network client of claim 1 , wherein the video decoder is implemented using the processor.
10. The network client of claim 1 , wherein the image processor is implemented using the processor.
1 1 . A method for augmented video transitions using a network client configured to receive streams of video data, comprising:
receiving a first stream of video data using a network client; playing the first stream of video data using the network client; detecting a change in streaming conditions using the network client;
requesting a second stream of video data using the network client; rendering an image for display using the network client;
receiving a second stream of video data using the network client; and playing the second stream of video data using the network client;
wherein the displayed image is displayed between a last frame of the first stream of video data displayed and a first frame of the second stream of video data displayed.
12. The method of claim 1 1 , further comprising receiving the image for display using the network client.
13. The method of claim 1 1 , wherein:
the first stream of video data includes a first group of pictures, where the first group of pictures comprises a first frame and a last frame; and
generating the image for display further comprises using the first group of pictures to generate the image for display.
14. The method of claim 13, wherein generating the image for display utilizes the last frame of the first group of pictures.
15. The method of claim 1 1 , wherein:
the second stream of video data includes a second group of pictures, where the second group of pictures comprises a first frame and a last frame; and
generating the image for display further comprises using the second group of pictures to generate the image for display.
16. The method of claim 15, wherein generating the image for display utilizes the first frame of the second group of pictures.
17. The method of claim 1 1 , wherein the image for display is pre-computed.
18. The method of claim 1 1 , further comprising determining the image for display in real time using the network client.
19. The method of claim 1 1 , further comprising resizing the image for display using the network client.
20. A machine readable medium containing processor instructions, where execution of the instructions by a processor causes the processor to perform a process comprising:
receiving a first stream of video data;
playing the first stream of video data;
detecting a change in streaming conditions;
requesting a second stream of video data;
rendering an image for display;
receiving a second stream of video data; and
playing the second stream of video data;
wherein the displayed image is displayed between a last frame of the first stream of video data displayed and a first frame of the second stream of video data displayed.
21 . A media server, comprising:
multimedia content storage containing multimedia content comprising a set of alternative streams of video data; and
a processor;
wherein the processor is configured to:
receive a request to stream multimedia content stored in the multimedia content storage;
stream a first stream of video data, where the first stream of video data is one of the set of alternative streams of video data contained in the multimedia content; receive a request for a second stream of video data, where the second stream of video data is one of the set of alternative streams of video data contained in the multimedia content;
transmit an image, where the image is transmitted in response to the request for the second stream of video data; and
stream the second stream of video data.
22. The media server of claim 21 , wherein:
the first stream of video data comprises a first group of pictures, where the first group of pictures comprises a first frame and a last frame; and
the processor is configured to generate the transmitted image using the first group of pictures.
23. The media server of claim 22, wherein the transmitted image is the last frame of the first group of pictures.
24. The media server of claim 21 , wherein:
the second stream of video data comprises a second group of pictures, where the second group of pictures comprises a first frame and a last frame; and
the processor is configured to generate the transmitted image using the second group of pictures.
25. The media server of claim 24, wherein the transmitted image is the first frame of the second group of pictures.
26. The media server of claim 21 , wherein the processor is configured to generate the transmitted image in response to the request for the second stream of video data.
27. The media server of claim 21 , wherein the processor is configured to retrieve a pre-computed image to use as the transmitted image in response to the request for the second stream of video data.
28. The media server of claim 27, wherein the processor is configured to retrieve the pre-computed image from an image server.
29. The media server of claim 21 , wherein the multimedia content is encoded using Scalable Video Coding.
30. The media server of claim 21 , wherein the multimedia content is stored in a Matroska container.
31 . A method for augmented video transitions using a media server comprising multimedia content, comprising:
receiving a request to stream multimedia content using a media server, where the multimedia content comprises a set of alternative streams of video data;
streaming a first stream of video data using the media server, where the first stream of video data is one of the set of alternative streams of video data;
receiving a request for a second stream of video data using the media server, where the second stream of video data is one of the set of alternative streams of video data;
transmitting an image to the network client using the media server, where the image is transmitted in response to the request for the second stream of video data; and
streaming the second stream of video data using the media server.
32. The method of claim 31 , wherein:
the first stream of video data includes a first group of pictures, where the first group of pictures comprises a first frame and a last frame; and
generating the image for display further comprises using the first group of pictures to generate the image for display.
33. The method of claim 32, wherein the transmitted image is the last frame of the first group of pictures.
34. The method of claim 31 , wherein:
the second stream of video data includes a second group of pictures, where the second group of pictures comprises a first frame and a last frame; and
generating the image for display further comprises using the second group of pictures to generate the image for display.
35. The method of claim 34, wherein the transmitted image is the first frame of the second group of pictures.
36. The method of claim 31 , further comprising generating the transmitted image in response to the request for the second stream of video data.
37. The method of claim 31 , further comprising retrieving the transmitted image in response to the request for the second stream of video data.
38. The method of claim 37, wherein the transmitted image is retrieved from an image server.
39. The method of claim 31 , wherein the multimedia content is encoded using Scalable Video Coding.
40. The method of claim 31 , wherein the multimedia content is stored in a Matroska container.
41 . A machine readable medium containing processor instructions, where execution of the instructions by a processor causes the processor to perform a process comprising:
streaming a first stream of video data;
receiving a request for a second stream of video data;
transmitting an image, where the image is transmitted in response to the request for the second stream of video data; and
streaming a second stream of video data.
PCT/US2013/037851 2012-04-27 2013-04-23 Systems and methods for adaptive streaming with augmented video stream transitions WO2013163221A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US13/457,911 US9060184B2 (en) 2012-04-27 2012-04-27 Systems and methods for adaptive streaming with augmented video stream transitions using a media server
US13/457,901 2012-04-27
US13/457,911 2012-04-27
US13/457,901 US20130287092A1 (en) 2012-04-27 2012-04-27 Systems and Methods for Adaptive Streaming with Augmented Video Stream Transitions

Publications (1)

Publication Number Publication Date
WO2013163221A1 true WO2013163221A1 (en) 2013-10-31

Family

ID=49483834

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2013/037851 WO2013163221A1 (en) 2012-04-27 2013-04-23 Systems and methods for adaptive streaming with augmented video stream transitions

Country Status (1)

Country Link
WO (1) WO2013163221A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9060184B2 (en) 2012-04-27 2015-06-16 Sonic Ip, Inc. Systems and methods for adaptive streaming with augmented video stream transitions using a media server
US10083621B2 (en) 2004-05-27 2018-09-25 Zedasoft, Inc. System and method for streaming video into a container-based architecture simulation

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6678332B1 (en) * 2000-01-04 2004-01-13 Emc Corporation Seamless splicing of encoded MPEG video and audio
US20060233259A1 (en) * 2005-04-15 2006-10-19 Chou Peter H Switching decode resolution during video decoding
US20110280153A1 (en) * 2010-05-13 2011-11-17 Futurewei Technologies, Inc. System, Apparatus for Content Delivery for Internet Traffic and Methods Thereof
US20120039582A1 (en) * 2009-04-20 2012-02-16 Koninklijke Philips Electronics N.V. Verification and synchronization of files obtained separately from a video content

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6678332B1 (en) * 2000-01-04 2004-01-13 Emc Corporation Seamless splicing of encoded MPEG video and audio
US20060233259A1 (en) * 2005-04-15 2006-10-19 Chou Peter H Switching decode resolution during video decoding
US20120039582A1 (en) * 2009-04-20 2012-02-16 Koninklijke Philips Electronics N.V. Verification and synchronization of files obtained separately from a video content
US20110280153A1 (en) * 2010-05-13 2011-11-17 Futurewei Technologies, Inc. System, Apparatus for Content Delivery for Internet Traffic and Methods Thereof

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10083621B2 (en) 2004-05-27 2018-09-25 Zedasoft, Inc. System and method for streaming video into a container-based architecture simulation
US9060184B2 (en) 2012-04-27 2015-06-16 Sonic Ip, Inc. Systems and methods for adaptive streaming with augmented video stream transitions using a media server

Similar Documents

Publication Publication Date Title
US11025902B2 (en) Systems and methods for the reuse of encoding information in encoding alternative streams of video data
US9350780B2 (en) Insertion of graphic overlays into a stream
US11546643B2 (en) Systems and methods for providing audio content during trick-play playback
US20180063590A1 (en) Systems and Methods for Encoding and Playing Back 360° View Video Content
US9060184B2 (en) Systems and methods for adaptive streaming with augmented video stream transitions using a media server
US9354799B2 (en) Systems and methods for adaptive streaming systems with interactive video timelines
EP2629476A1 (en) Adaptive display streams
KR20200109359A (en) Video streaming
AU2018250308B2 (en) Video compression using down-sampling patterns in two phases
WO2013163221A1 (en) Systems and methods for adaptive streaming with augmented video stream transitions
US20130287092A1 (en) Systems and Methods for Adaptive Streaming with Augmented Video Stream Transitions
US20140181653A1 (en) Content presentation with enhanced user experience
US9160933B2 (en) Luminance based image capturing methods and apparatus

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 13781835

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 13781835

Country of ref document: EP

Kind code of ref document: A1