US6044397A - System and method for generation and interfacing of bitstreams representing MPEG-coded audiovisual objects - Google Patents

System and method for generation and interfacing of bitstreams representing MPEG-coded audiovisual objects Download PDF

Info

Publication number
US6044397A
US6044397A US09/055,929 US5592998A US6044397A US 6044397 A US6044397 A US 6044397A US 5592998 A US5592998 A US 5592998A US 6044397 A US6044397 A US 6044397A
Authority
US
United States
Prior art keywords
audiovisual
bitstream
streaming
interface library
bitstreams
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
US09/055,929
Inventor
Alexandros Eleftheriadis
Yihan Fang
Atul Puri
Robert Lewis Schmidt
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Columbia University of New York
AT&T Corp
Original Assignee
Columbia University of New York
AT&T Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Columbia University of New York, AT&T Corp filed Critical Columbia University of New York
Priority to US09/055,929 priority Critical patent/US6044397A/en
Assigned to AT&T CORP. reassignment AT&T CORP. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: PURI, ATUL, SCHMIDT, ROBERT LEWIS
Assigned to COLUMBIA UNIVERSITY reassignment COLUMBIA UNIVERSITY ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: FANG, YIHAN, ELEFTHERIADIS, ALEXANDROS
Application granted granted Critical
Publication of US6044397A publication Critical patent/US6044397A/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/23412Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs for generating or manipulating the scene composition of objects, e.g. MPEG-4 objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/20Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
    • H04N19/25Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding with scene description coding, e.g. binary format for scenes [BIFS] compression
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/20Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
    • H04N19/27Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding involving both synthetic and natural picture components, e.g. synthetic natural hybrid coding [SNHC]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs
    • H04N21/44012Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream, rendering scenes according to MPEG-4 scene graphs involving rendering scenes according to scene graphs, e.g. MPEG-4 scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/443OS processes, e.g. booting an STB, implementing a Java virtual machine in an STB or power management in an STB
    • H04N21/4431OS processes, e.g. booting an STB, implementing a Java virtual machine in an STB or power management in an STB characterized by the use of Application Program Interface [API] libraries
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/24Systems for the transmission of television signals using pulse code modulation
    • H04N7/52Systems for transmission of a pulse code modulated video signal with one or more other pulse code modulated signals, e.g. an audio signal or a synchronizing signal

Definitions

  • the invention relates to the field of coded multimedia and its storage, transmission and delivery to users, and more particularly to such coding when a flexible means for generating, editing or interpreting bitstreams representing multimedia objects is necessary.
  • Digital multimedia offers advantages of manipulation, multigeneration processing and error robustness and others, but incurs constraints due to the storage capacity or transmission bandwidth required. Multimedia content thus frequently needs to be compressed or coded. Further, in the wake of rapid increases in demand for digital multimedia over the Internet and other networks, the need for efficient storage, networked access and search and retrieval has increased, and a number of coding schemes, storage formats, retrieval techniques and transmission protocols have evolved. For instance, for image and graphics files, GIF, TIF and other formats have been used. Similarly, audio files have been coded and stored in RealAudio, WAV, MIDI and other formats. Animations and video files have often been stored using GIF89a, Cinepak, Indeo and others.
  • decoders and interpreters are often needed, and may offer various degrees of speed, quality and performance depending on whether these decoders and interpreters are implemented in hardware or in software, and particularly in the case of software, on the capabilities of the host computer. If multimedia content is embedded in web pages accessed via a computer (e.g. a PC), the web browser needs to be set up correctly for all the anticipated content and must recognize each type of content and support a mechanism of content handlers (software plugins or hardware) to deal with such content.
  • a computer e.g. a PC
  • ISO MPEG International Standards Organization Motion Pictures Experts Group
  • DSM digital storage media
  • MPEG-2 was primarily intended for use in broadcast environment (transport stream), although it also supports MPEG- 1 like mechanism for use on DSM (program stream).
  • MPEG-2 also included additional features such as DSM-Control and Command for basic user interaction, as may be needed for standardized playback of MPEG-2, either standalone or networked.
  • DSM-Control and Command for basic user interaction
  • the MPEG-1 standard is becoming commonly available for playback of movies and games on PCs.
  • the MPEG-2 standard since it addresses relatively higher quality applications, is becoming common for entertainment applications via digital satellite TV, digital cable and Digital Versatile Disk (DVD).
  • MPEG-1 and MPEG-2 are expected to be utilized in various other configurations, in streams communicated over network, streams stored on hard disks or CDs, and in the combination of networked and local access.
  • MPEG-4 a new standard
  • MPEG-4 has addressed coding of audio-visual information in the form of individual objects, and a system for composition and synchronized playback of these objects.
  • the Java language offers networking and platform independence critical to downloading and executing of applets (java classes) on a client PC from a web server which hosts the web pages visited by the user.
  • applets java classes
  • a single access to the data stored on the server may be needed and all the necessary data may be stored on the client PC, or several partial accesses (to reduce storage space and time needed for startup) may be needed.
  • the later scenario is referred to as streamed playback.
  • the bandwidth for networked access of multimedia may be either limited or time-varying, necessitating transmission of the most significant information only and followed by other information as more bandwidth becomes available.
  • the client side PC on which decoding may have to take place may be limited in CPU and/or memory resources, and furthermore, these resources may also be time-varying.
  • a multimedia user (consumer) may require highly interactive nonlinear browsing and playback.
  • APIs application programming interfaces
  • MPEG-4 has used a binary format screen representation (BIFS) of the constructs central to VRML and extended VRML in many ways to handle real-time audio/video data and effects such as facial or body animation. Since the MPEG-4 standard offers many tools for coding of various types of media as well as scene graph representation, and further, each media coding may involve separate coding of individual objects, an organized yet flexible mechanism for bitstream generation, editing and interpretation is highly desirable.
  • the invention addresses standardized interfaces for MPEG-4 authoring, bitstream manipulation, editing and interpretation.
  • the invention provides tools and interfaces to considerably facilitate the aforementioned operations, resulting in coded bitstreams which are easier to test, check and debug while conforming to the MPEG-4 standard.
  • the specified interfaces can also facilitate graceful adaptation by allowing editing of bitstreams when sufficient processing resources may not exist.
  • the specified interfaces can also allow creation of decodable bitstreams in response to the user request either directly or indirectly embedded in audiovisual applications and services expected to be important in near future.
  • the current invention not only addresses the shortcomings of a conventional system of bitstream encoding and decoding, but also offers tools which can be integrated into more adaptive systems such as that supporting graceful degradation as well as responsiveness to user interaction.
  • the invention provides a system and interfacing method facilitating flexible generation, editing and interpretation of bitstreams representing audio-visual objects coded by the MPEG-4 standard.
  • the invention specifies a bitstream input/output package in Java programming language to facilitates bitstream encoding and decoding of audio-visual media objects, especially when coding is per the MPEG-4 standard. This is due to the fact that the proposed package separates fixed length and variable length coding and allows flexible parsing which offers the potential of optimized implementation as needed to aid real-time or near real-time operation.
  • the invention is motivated in part by the desirability of standardized interfaces for MPEG-4 authoring, bitstream manipulation, editing and interpretation.
  • One goal of the invention is to provide tools and interfaces to considerably facilitate the aforementioned audiovisual operations, resulting in coded bitstreams which are easier to test, check and debug while conforming to the MPEG-4 standard.
  • the interfaces specified in the invention can also facilitate the graceful adaptation by allowing editing of bitstreams when sufficient processing resources may not exist.
  • the specified interfaces can also allow creation of decodable bitstreams in response to the user request either directly or indirectly embedded in audiovisual applications and services expected to be important in the future.
  • FIG. 1A illustrates a block diagram of the encoding system with details illustrating an embodiment of the invention
  • FIG. 1B illustrates a block diagram of the decoding system with details illustrating the embodiment of the invention
  • FIG. 2 illustrates bitstream generation interface according to the invention
  • FIG. 3 illustrates bitstream editing and interpretation interface according to the invention.
  • FIG. 4 illustrates a flowchart depicting a buffer updating process used in the invention.
  • the invention provides an integrated interface facility for processing streamed audiovisual information, and will be illustratively described in the MPEG-4 environment.
  • the interface facility provided by the invention includes a bitstream input/output library for flexible generation, editing and interpretation of bitstreams representing individual audiovisual objects coded using the MPEG-4 standard.
  • the invention defines a bitstream input/output package in the Java language. This package, mpgj.bitsio, can be added to the standard Java libraries known in the art, and simplify the bitstream Input and Output operations, including fixed length and variable length coding common in MPEG-4 syntax decoding.
  • the package is organized to facilitate optimization for maximum speed.
  • the variable length code parsing module could use configurable multi-stage parallel lookups to further aid real-time or near real-time operation.
  • the library is the Java equivalent of bits input/output portion of MSDL-S (MPEG-4 Syntactic Description Language) known in the art.
  • MSDL-S MPEG-4 Syntactic Description Language
  • the invention can also be used internally by the MSDL-S to Java translator (flavorj).
  • FIG. 1A illustrates a block diagram of an encoding system with details illustrating an embodiment of the invention.
  • a natural source video scene to be coded is input via line 100 to Video Segmenter 101, which segments the scene into a number of semantic objects and are output on lines 103,104,105. Other video objects external to be scene can also be mixed such as that on line 102.
  • video objects 102,103,104,105 sequentially pass through the switch 112 and are available one after the other on line 113, an input to Media1 Encoder 118.
  • a natural source audio scene to be coded on line 138 is segmented by Audio Segmenter 106 into individual objects, 108, 109, 110, and any external objects 107.
  • audio objects 107,108,109,110 sequentially pass through switch 114 and are available one after the other on line 115, an input to Media2 Encoder.
  • synthetic objects either aural or visual are input to line 111, an input to Media3 Encoder, 120.
  • a description of scene is also generated optionally in Scene Graph 117, which is output on line 124.
  • SceneGraph 117 also generates an optional control signal which is sent to each of the media encoders, for example to Media1 Encoder 118 over line 121, to Media2 Encoder 119 over line 122 and Media3 Encoder 120 over line 127.
  • media encoders there is no limit on the number of media encoders that can be employed according to the invention. Further, it is possible that a media encoder may itself consist of subencoders.
  • the output of media encoders on lines 125,126 and 127 forms input to respective media bits(tream) generators, Media1 Bits Generator 129, Media2 Bits Generator 130 and Media3 Bits Generator 131.
  • the output of Scene Graph 117 on line 124 forms input to BIFS Bits Generator 128.
  • BIFS Bits Generator 128 as well as Media Bits Generators 129,130 and 131 are assumed to employ interfaces of this invention detailed in FIG. 2.
  • the output of various bits generators on lines 132, 133 and 134 and 135 are fed to systems multiplexer, Mux, 136.
  • the multiplexed bitstream is available on Channel 137 for storage or transmission.
  • FIG. 1B illustrates a block diagram of the decoding operation of the invention, in more detail. Besides a few exceptions, the operation of this aspect of the invention is the inverse of that of the coding depicted in FIG. 1A.
  • the multiplexed bitstream (either from storage or transmission) is available on Channel 137 and is input to demultiplexer Demux 151, which separates that stream into individual bitstreams, such as visual (natural and synthetic video objects), audio (natural and synthetic audio objects), BIFS scene description etc.
  • the BIFS scene bitstream is available on line 152, video objects bitstream on line 153, audio objects bitstream on line 154 and synthetic (visual or audio) objects bitstream on line 155 and forms inputs to the Bits(tream) Editor 156, which responds to the several conditions requiring graceful degradation or functionalities on user request.
  • the Bits Editor 156 can be used for real-time as well as non realtime editing of bitstreams, and employs interfaces of this invention, as detailed in FIG. 3.
  • the modified bitstreams, BIFS Bits on line 157, Media 1 Bits on line 158, Media 2 Bits on line 159, Media3 Bits on line 160 are input to the respective bitstream interpreters, BIFS Bits Interpreter 161, Media1 Bits Interpreter 162, Media2 Bits Interpreter 163, Media3 Bits Interpreter 164 which output respective streams of symbols on lines 165, 166, 167 and 168.
  • the bitstream interpreters use interfaces of FIG.
  • the BIFS symbols on line 165 are decoded to form a Scene Graph on line 179, various types of Media symbol streams on lines 166, 167, 168 are decoded by respective media decoders, Media1 Decoder 170, Media2 Decoder 171 and Media3 Decoder 172 and the decoded media streams (video objects, audio objects, synthetic objects etc.) are output on lines 176,177 and 178.
  • the various media decoders are instantiated by Scene Graph 169 and controls to these decoders are shown; Media1 Decoder is controlled via line 173, Media2 Decoder is controlled via line 174 and Media3 Decoder is controlled via line 175.
  • a Comp(ositor) 182 takes as input the Scene Graph on line 180, and the output of the three media decoders on lines 176,177 and 178 and composes a scene to be presented to viewer/user, but first, the output of the compositor on line 183 is transmitted to the renderer which is also controlled by the Scene Graph on line 181, and renders the composited scene.
  • FIG. 2 illustrates the bitstream generation interface according to the invention.
  • the BIFS Bits Generator 128 takes as input BIFS symbols on line 124 and outputs corresponding coded representation in the form of bitstream on line 132.
  • Media1,2,3 Bits Generators, 129,130,131 take as input media symbols on lines 125,126 and 127 and output the corresponding coded representation in the form of bitstreams on lines 133,134 and 135 respectively.
  • the BIFS Bits Generator 128 as well as Media1,2,3 Bits Generator 129,130,131 employ Bits Generator Interface, 200 of the invention.
  • Bits Generator Interface 200 consists of Java classes such as OutputStream, 201, and Map, 202. The interface operation of these classes is described as follows.
  • OutputStream is the basic interface to the output stream.
  • a map table for fixed or variable length coding is a map table for fixed or variable length coding.
  • FIG. 3 illustrates a bitstream editing and interpretation interface according to the invention.
  • the BIFS Bits Interpreter 161 takes as input BIFS bitstream on line 157 and outputs corresponding decoded symbols on line 165.
  • Media1,2,3 Bits Interpreters, 162,163,164 take as input media bitstreams on lines 158,159 and 160 and output the corresponding decoded symbols on lines 166,167 and 168 respectively.
  • the BIFS Bits Interpreter 161 as well as Media1,2,3 Bits Interpreter 162,163,164 employ Bits Editor/Interpreter Interface, 300 of this invention.
  • Bits Editor/Interpreter Interface 300 consists of Java classes such as InputStream 303, Map 302, and OutputStream, 301. Besides bitstream interpretation, the interfaces of this figure also support bitstream editing.
  • Bits Editor 156 takes as input the demultiplexed BIFS and media bitstreams on lines 152,153,154, 155 and outputs the corresponding modified (edited bitstreams) on respective lines 157,158,159,160.
  • the editing operation can be in response to the need to discard objects due to overloaded system resources, or to that of user interaction, and is enabled by BIFS Editor/Interpreter Interface, 300.
  • this interface supports InputStream and Map classes, however the interface also supports OutputStream class, the later (along with Map) needed for bitstream editing operations.
  • the interface operation of OutputStream and Map classes were discussed earlier (in conjunction with FIG. 2).
  • the interface operation of InputStream class is as follows.
  • This class is the basic interface to the input stream.
  • Probe next specified number of bits Return value as 32-bit integer. Do not advance current pointer. Sets error flag if stream cannot be read.
  • Probe variable or fixed length code according to specified vlc map. Return value as 32-bit integer. Sets eos flag on end-of-stream. Sets error flag if stream cannot be read.
  • FIG. 4 An aspect of the foregoing routines, generally related to processing given lengths of bit strings using a data buffer, is illustrated in FIG. 4.
  • an input length is read at 400, after which a data buffer is checked at step 410.
  • the buffer may be filled (415) or read out (420) according to bitlength or other parameters, after which the buffer is updated (step 430).
  • a flexible bitstream facility is introduced, establishing core routines to enable simpler as well as more complicated control of embedded audiovisual objects, all in a universal and consistent streamed manner.

Abstract

The invention provides a standardized interface facility for MPEG-4 authoring, bitstream manipulation, editing and interpretation, with associated tools and interfaces to, resulting in coded bitstreams which are easier to test, check and debug while conforming to the MPEG-4 standard. The specified interfaces can also facilitate graceful degradation in the face of decreased resources by allowing editing of bitstreams. The specified interfaces can also allow creation of decodable bitstreams in response to the user requests either directly or indirectly embedded in audiovisual applications, as well as future services. The invention specifies a bitstream input/output package in the Java programming language to facilitates bitstream encoding and decoding of audio-visual media objects, especially when coding uses the MPEG-4 standard. The invention separates fixed length and variable length coding, and allows flexible parsing which offers the potential of optimized implementation as needed to aid real-time or near real-time operation.

Description

CROSS-REFERENCE TO RELATED APPLICATION
This application is related to U.S. Provisional Application Serial No. 60/042,801, from which priority is claimed.
BACKGROUND OF THE INVENTION
1. Field of Invention
The invention relates to the field of coded multimedia and its storage, transmission and delivery to users, and more particularly to such coding when a flexible means for generating, editing or interpreting bitstreams representing multimedia objects is necessary.
2. Description of Related Art
Digital multimedia offers advantages of manipulation, multigeneration processing and error robustness and others, but incurs constraints due to the storage capacity or transmission bandwidth required. Multimedia content thus frequently needs to be compressed or coded. Further, in the wake of rapid increases in demand for digital multimedia over the Internet and other networks, the need for efficient storage, networked access and search and retrieval has increased, and a number of coding schemes, storage formats, retrieval techniques and transmission protocols have evolved. For instance, for image and graphics files, GIF, TIF and other formats have been used. Similarly, audio files have been coded and stored in RealAudio, WAV, MIDI and other formats. Animations and video files have often been stored using GIF89a, Cinepak, Indeo and others. To play back the plethora of existing formats, decoders and interpreters are often needed, and may offer various degrees of speed, quality and performance depending on whether these decoders and interpreters are implemented in hardware or in software, and particularly in the case of software, on the capabilities of the host computer. If multimedia content is embedded in web pages accessed via a computer (e.g. a PC), the web browser needs to be set up correctly for all the anticipated content and must recognize each type of content and support a mechanism of content handlers (software plugins or hardware) to deal with such content.
The need for interoperability, guaranteed quality and performance and economies of scale in chip design, as well as the cost involved in content generation for multiplicity of formats, has led to advances in standardization in the areas of multimedia coding, packetization and robust delivery. In particular, International Standards Organization Motion Pictures Experts Group (ISO MPEG) has standardized bitstream syntax and decoding semantics for coded multimedia in the form of two standards, referred to as MPEG-1 and MPEG-2. MPEG-1 was primarily intended for use on digital storage media (DSM) such as compact disks (CDs), whereas MPEG-2 was primarily intended for use in broadcast environment (transport stream), although it also supports MPEG- 1 like mechanism for use on DSM (program stream). MPEG-2 also included additional features such as DSM-Control and Command for basic user interaction, as may be needed for standardized playback of MPEG-2, either standalone or networked. With the advent of inexpensive boards and PCMCIA cards and the availability of fast Central Processing Units (CPUs), the MPEG-1 standard is becoming commonly available for playback of movies and games on PCs. The MPEG-2 standard, on the other hand, since it addresses relatively higher quality applications, is becoming common for entertainment applications via digital satellite TV, digital cable and Digital Versatile Disk (DVD). Besides the applications/platforms noted, MPEG-1 and MPEG-2 are expected to be utilized in various other configurations, in streams communicated over network, streams stored on hard disks or CDs, and in the combination of networked and local access.
The success of MPEG-1 and MPEG-2, the bandwidth limitations of Internet and mobile channels, the flexibility of web based data access using browsers, and the increasing need for interactive personal communication has opened up new paradigms for multimedia usage and control. In response, ISO-MPEG has developed a new standard, called MPEG-4. The MPEG-4 standard has addressed coding of audio-visual information in the form of individual objects, and a system for composition and synchronized playback of these objects. While development of MPEG-4 for such fixed systems continues, in the meantime new paradigms in communication, software and networking such as that offered by the Java language have offered new opportunities for flexibility, adaptivity and user interaction. For instance, the Java language offers networking and platform independence critical to downloading and executing of applets (java classes) on a client PC from a web server which hosts the web pages visited by the user. Depending on the design of the applet, either a single access to the data stored on the server may be needed and all the necessary data may be stored on the client PC, or several partial accesses (to reduce storage space and time needed for startup) may be needed. The later scenario is referred to as streamed playback.
As noted, when coded multimedia is used for Internet and local networked applications on a computer, say a PC, a number of situations may arise. First, the bandwidth for networked access of multimedia may be either limited or time-varying, necessitating transmission of the most significant information only and followed by other information as more bandwidth becomes available. Second, regardless of the bandwidth available, the client side PC on which decoding may have to take place may be limited in CPU and/or memory resources, and furthermore, these resources may also be time-varying. Third, a multimedia user (consumer) may require highly interactive nonlinear browsing and playback. This is not unusual since a lot of textual content on web pages is capable of being browsed via use of hyperlinked features, and the same paradigm is expected for presentations employing coded audio-visual objects. The MPEG-4 system without enhanced capabilities may only be able to deal with the aforementioned situations in a very limited way.
The use of application programming interfaces (APIs) has long been recognized in the software industry as a means to achieve standardized operations and functions over a number of different types of computer platforms. Typically, although operations can be standardized via definition of API, the performance of these operations may differ on various platforms as specific vendors with interest in a specific platform may provide implementations optimized for that platform. In the field of graphics, Virtual Reality Modeling Language (VRML) allows a means of specifying spatial and temporal relationships between objects and description of a scene by use of a scene graph approach. MPEG-4 has used a binary format screen representation (BIFS) of the constructs central to VRML and extended VRML in many ways to handle real-time audio/video data and effects such as facial or body animation. Since the MPEG-4 standard offers many tools for coding of various types of media as well as scene graph representation, and further, each media coding may involve separate coding of individual objects, an organized yet flexible mechanism for bitstream generation, editing and interpretation is highly desirable.
SUMMARY OF THE INVENTION
The invention addresses standardized interfaces for MPEG-4 authoring, bitstream manipulation, editing and interpretation. The invention provides tools and interfaces to considerably facilitate the aforementioned operations, resulting in coded bitstreams which are easier to test, check and debug while conforming to the MPEG-4 standard. The specified interfaces can also facilitate graceful adaptation by allowing editing of bitstreams when sufficient processing resources may not exist. The specified interfaces can also allow creation of decodable bitstreams in response to the user request either directly or indirectly embedded in audiovisual applications and services expected to be important in near future. Thus, the current invention, not only addresses the shortcomings of a conventional system of bitstream encoding and decoding, but also offers tools which can be integrated into more adaptive systems such as that supporting graceful degradation as well as responsiveness to user interaction.
Specifically, the invention provides a system and interfacing method facilitating flexible generation, editing and interpretation of bitstreams representing audio-visual objects coded by the MPEG-4 standard. The invention specifies a bitstream input/output package in Java programming language to facilitates bitstream encoding and decoding of audio-visual media objects, especially when coding is per the MPEG-4 standard. This is due to the fact that the proposed package separates fixed length and variable length coding and allows flexible parsing which offers the potential of optimized implementation as needed to aid real-time or near real-time operation.
The invention is motivated in part by the desirability of standardized interfaces for MPEG-4 authoring, bitstream manipulation, editing and interpretation. One goal of the invention is to provide tools and interfaces to considerably facilitate the aforementioned audiovisual operations, resulting in coded bitstreams which are easier to test, check and debug while conforming to the MPEG-4 standard. The interfaces specified in the invention can also facilitate the graceful adaptation by allowing editing of bitstreams when sufficient processing resources may not exist. The specified interfaces can also allow creation of decodable bitstreams in response to the user request either directly or indirectly embedded in audiovisual applications and services expected to be important in the future.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention will be described with reference to the accompanying drawings, in which like elements are designated by like numbers and in which:
FIG. 1A illustrates a block diagram of the encoding system with details illustrating an embodiment of the invention;
FIG. 1B illustrates a block diagram of the decoding system with details illustrating the embodiment of the invention;
FIG. 2 illustrates bitstream generation interface according to the invention;
FIG. 3 illustrates bitstream editing and interpretation interface according to the invention; and
FIG. 4 illustrates a flowchart depicting a buffer updating process used in the invention.
DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
The invention provides an integrated interface facility for processing streamed audiovisual information, and will be illustratively described in the MPEG-4 environment. The interface facility provided by the invention includes a bitstream input/output library for flexible generation, editing and interpretation of bitstreams representing individual audiovisual objects coded using the MPEG-4 standard. In one aspect the invention defines a bitstream input/output package in the Java language. This package, mpgj.bitsio, can be added to the standard Java libraries known in the art, and simplify the bitstream Input and Output operations, including fixed length and variable length coding common in MPEG-4 syntax decoding. Among other advantages, the package is organized to facilitate optimization for maximum speed. For example, the variable length code parsing module could use configurable multi-stage parallel lookups to further aid real-time or near real-time operation.
              TABLE 1                                                     
______________________________________                                    
The bitstream input/output library of invention                           
#   Classes    Explanation                                                
______________________________________                                    
1.  InputStream                                                           
               This class provides the bitstream input capabilities       
2.      Map                This class is used by the InputStream and      
               OutputStream classes                                       
3.      OutputStream                                                      
                  This class provides the bitstream output                
______________________________________                                    
               capabilities                                               
The library is the Java equivalent of bits input/output portion of MSDL-S (MPEG-4 Syntactic Description Language) known in the art. In fact, the invention can also be used internally by the MSDL-S to Java translator (flavorj).
FIG. 1A illustrates a block diagram of an encoding system with details illustrating an embodiment of the invention. A natural source video scene to be coded is input via line 100 to Video Segmenter 101, which segments the scene into a number of semantic objects and are output on lines 103,104,105. Other video objects external to be scene can also be mixed such as that on line 102. Next, video objects 102,103,104,105, sequentially pass through the switch 112 and are available one after the other on line 113, an input to Media1 Encoder 118. In parallel, a natural source audio scene to be coded on line 138 is segmented by Audio Segmenter 106 into individual objects, 108, 109, 110, and any external objects 107. Next, audio objects 107,108,109,110 sequentially pass through switch 114 and are available one after the other on line 115, an input to Media2 Encoder. Besides natural audio and video objects, synthetic objects, either aural or visual are input to line 111, an input to Media3 Encoder, 120. Based on content authors input, 116, a description of scene is also generated optionally in Scene Graph 117, which is output on line 124. SceneGraph 117 also generates an optional control signal which is sent to each of the media encoders, for example to Media1 Encoder 118 over line 121, to Media2 Encoder 119 over line 122 and Media3 Encoder 120 over line 127.
Although three media encoders are shown, there is no limit on the number of media encoders that can be employed according to the invention. Further, it is possible that a media encoder may itself consist of subencoders. The output of media encoders on lines 125,126 and 127 forms input to respective media bits(tream) generators, Media1 Bits Generator 129, Media2 Bits Generator 130 and Media3 Bits Generator 131. The output of Scene Graph 117 on line 124 forms input to BIFS Bits Generator 128. BIFS Bits Generator 128 as well as Media Bits Generators 129,130 and 131 are assumed to employ interfaces of this invention detailed in FIG. 2. The output of various bits generators on lines 132, 133 and 134 and 135 are fed to systems multiplexer, Mux, 136. The multiplexed bitstream is available on Channel 137 for storage or transmission.
FIG. 1B illustrates a block diagram of the decoding operation of the invention, in more detail. Besides a few exceptions, the operation of this aspect of the invention is the inverse of that of the coding depicted in FIG. 1A. The multiplexed bitstream (either from storage or transmission) is available on Channel 137 and is input to demultiplexer Demux 151, which separates that stream into individual bitstreams, such as visual (natural and synthetic video objects), audio (natural and synthetic audio objects), BIFS scene description etc. The BIFS scene bitstream is available on line 152, video objects bitstream on line 153, audio objects bitstream on line 154 and synthetic (visual or audio) objects bitstream on line 155 and forms inputs to the Bits(tream) Editor 156, which responds to the several conditions requiring graceful degradation or functionalities on user request. The Bits Editor 156 can be used for real-time as well as non realtime editing of bitstreams, and employs interfaces of this invention, as detailed in FIG. 3.
The modified bitstreams, BIFS Bits on line 157, Media 1 Bits on line 158, Media 2 Bits on line 159, Media3 Bits on line 160 are input to the respective bitstream interpreters, BIFS Bits Interpreter 161, Media1 Bits Interpreter 162, Media2 Bits Interpreter 163, Media3 Bits Interpreter 164 which output respective streams of symbols on lines 165, 166, 167 and 168. The bitstream interpreters use interfaces of FIG. 3 of the invention The BIFS symbols on line 165 are decoded to form a Scene Graph on line 179, various types of Media symbol streams on lines 166, 167, 168 are decoded by respective media decoders, Media1 Decoder 170, Media2 Decoder 171 and Media3 Decoder 172 and the decoded media streams (video objects, audio objects, synthetic objects etc.) are output on lines 176,177 and 178. The various media decoders are instantiated by Scene Graph 169 and controls to these decoders are shown; Media1 Decoder is controlled via line 173, Media2 Decoder is controlled via line 174 and Media3 Decoder is controlled via line 175. A Comp(ositor) 182 takes as input the Scene Graph on line 180, and the output of the three media decoders on lines 176,177 and 178 and composes a scene to be presented to viewer/user, but first, the output of the compositor on line 183 is transmitted to the renderer which is also controlled by the Scene Graph on line 181, and renders the composited scene.
FIG. 2 illustrates the bitstream generation interface according to the invention. The BIFS Bits Generator 128 takes as input BIFS symbols on line 124 and outputs corresponding coded representation in the form of bitstream on line 132. Similarly, Media1,2,3 Bits Generators, 129,130,131 take as input media symbols on lines 125,126 and 127 and output the corresponding coded representation in the form of bitstreams on lines 133,134 and 135 respectively. The BIFS Bits Generator 128 as well as Media1,2,3 Bits Generator 129,130,131 employ Bits Generator Interface, 200 of the invention. Bits Generator Interface 200 consists of Java classes such as OutputStream, 201, and Map, 202. The interface operation of these classes is described as follows.
Class mpgj.bitsio.OutputStream
java.lang.Object
|
+ - - - mpgj.bitsio.OutputStream
public class OutputStream extends Object
OutputStream is the basic interface to the output stream.
Constructors
public OutputStream(FileOutputStream file)
Constructs a new OutputStream in a file.
public OutputStream(string bits)
Constructs a new OutputStream in a string.
Methods
public void align(int numbits)
Align with next bit boundary that is a multiple of numbits. Bits between the current pointer and the alignment boundary are written out as zeros.
public void align(string stuffing[ ], int numbits)
Align with next bit boundary that is a multiple of numbits. Bits between the current pointer and the alignment boundary are stuffed according to the stuffing string.
public boolean eos()
Return true on end-of-stream or error, else return false.
public boolean error()
Return true on error, else return false.
public void putbits(int numbits, int value)
Put an unsigned integer using the specified number of bits. Sets error flag if stream cannot be written or value is negative.
public void putsbits(int numbits, int value)
Put a signed integer using the specified number of bits. The last bit is set to be sign bit. Sets error flag if stream cannot be written.
public void putvlc(Map map, int value)
Put specified value using the specified map. Sets error flag if stream cannot be written.
Class mpgj.bitsio.Map
java.lang.Object
|
+ - - - mpgj.bitsio.Map
public class Map extends Object
A map table for fixed or variable length coding.
Constructors
public Map(FileInputStream file, int step)
Constructs a new map reading from a file, using specified step for lookups.
public Map(FilelnputStream file)
Constructs a new map reading from a file. Every lookup advances 1 bit.
public Map(string[ ] bitstring, int[ ] value, int array-- size, int step)
Constructs a new map from a string array and an integer array knowing array size and specified step.
public Map(string[ ] bitstring, int[ ] value, int array-- size)
Constructs a new map from a string array and an integer array knowing array size. Every lookup advances 1 bit.
Methods
No user-level methods are provided.
FIG. 3 illustrates a bitstream editing and interpretation interface according to the invention. The BIFS Bits Interpreter 161 takes as input BIFS bitstream on line 157 and outputs corresponding decoded symbols on line 165. Similarly, Media1,2,3 Bits Interpreters, 162,163,164 take as input media bitstreams on lines 158,159 and 160 and output the corresponding decoded symbols on lines 166,167 and 168 respectively. The BIFS Bits Interpreter 161 as well as Media1,2,3 Bits Interpreter 162,163,164 employ Bits Editor/Interpreter Interface, 300 of this invention. Bits Editor/Interpreter Interface 300 consists of Java classes such as InputStream 303, Map 302, and OutputStream, 301. Besides bitstream interpretation, the interfaces of this figure also support bitstream editing.
The bitstream editing operations are in general similar to the bitstream generation process. For instance, Bits Editor 156 takes as input the demultiplexed BIFS and media bitstreams on lines 152,153,154, 155 and outputs the corresponding modified (edited bitstreams) on respective lines 157,158,159,160. The editing operation can be in response to the need to discard objects due to overloaded system resources, or to that of user interaction, and is enabled by BIFS Editor/Interpreter Interface, 300. As noted, this interface supports InputStream and Map classes, however the interface also supports OutputStream class, the later (along with Map) needed for bitstream editing operations. The interface operation of OutputStream and Map classes were discussed earlier (in conjunction with FIG. 2). The interface operation of InputStream class is as follows.
Class mpgj.bitsio.InputStream
java.lang.Object
|
+ - - - mpgj.bitsio.InputStream
public class Input Stream extends Object
This class is the basic interface to the input stream.
Constructors
public InputStream(FilelnputStream file)
Constructs a new InputStream from a file.
public InputStream(string bits, int length)
Constructs a new InputStream from a string with given length.
Methods
public void align(int numbits)
Align with next bit boundary that is a multiple of numbits. Bits between the current pointer and the alignment boundary are read in and discarded.
public boolean eos()
Return true on end-of-stream, else return false.
public boolean error()
Return true on error, else return false.
public int getbits(int length)
Get an unsigned integer from specified number of bits. Sets eos flag on end-of-stream. Sets error flag if stream cannot be read.
public int getsbits(int length)
Get a signed integer from specified number of bits (length-1). The last bit indicates the sign of the integer. Sets eos flag on end-of-stream. Sets error flag if stream cannot be read.
public int nextbits(int length)
Probe next specified number of bits. Return value as 32-bit integer. Do not advance current pointer. Sets error flag if stream cannot be read.
public void skipbits(int length)
Skip specified number of bits. Sets eos flag on end-of-stream.
public int getvlc(Map map)
Get variable or fixed length code according to specified vlc map. Return value as 32-bit integer. Sets eos flag on end-of-stream. Sets error flag if stream cannot be read.
public int nextvlc(Map map)
Probe variable or fixed length code according to specified vlc map. Return value as 32-bit integer. Sets eos flag on end-of-stream. Sets error flag if stream cannot be read.
An aspect of the foregoing routines, generally related to processing given lengths of bit strings using a data buffer, is illustrated in FIG. 4. In that figure, an input length is read at 400, after which a data buffer is checked at step 410. According to the desired operation, the buffer may be filled (415) or read out (420) according to bitlength or other parameters, after which the buffer is updated (step 430).
In the implementation of the invention, a flexible bitstream facility is introduced, establishing core routines to enable simpler as well as more complicated control of embedded audiovisual objects, all in a universal and consistent streamed manner.
The foregoing description of the system and method of the invention is illustrative, and variations in construction and implementation will occur to persons skilled in the art. For instance, while an illustrative set of streaming functions have been described, functions can be added or subtracted according to changing network, application or other needs. The scope of the invention is intended to be limited only by the following claims.

Claims (20)

What is claimed is:
1. A system for processing streaming audiovisual objects coded according to the MPEG-4 standard, comprising:
a streaming interface library containing a predetermined set of streaming control functions processing audiovisual objects, each of the control functions having predefined function calls; and
a processor, configured to access the streaming interface library, and to decode and process streamed audiovisual objects according to the function calls.
2. The system of claim 1, wherein the processor unit executes a client application invoking the function calls.
3. The system of claim 1, further comprising a user input unit, the user input unit communicating with the processor unit and invoking selected function calls.
4. The system of claim 1, wherein the interface library comprises a visual decoding interface to decode visual objects in audiovisual bitstreams.
5. The system of claim 1, wherein the streaming interface library comprises a generation function for generating and outputting audiovisual data streams.
6. The system of claim 1, wherein the streaming interface library comprises an editing function for editing and manipulating audiovisual bitstreams.
7. The system of claim 1, wherein the streaming interface library comprises an interpreting function for interpreting audiovisual bitstreams.
8. The system of claim 1, wherein the processor unit adapts the execution of the bitstream interface library according to varying system resources.
9. The system of claim 1, further comprising a client application interface, the client application interface invoking additional bitstream functions presented by cooperating client applications.
10. The system of claim 1, further comprising a multimedia browser module employing the bitstream interface library for user viewing.
11. A method for processing streaming audiovisual objects coded according to the MPEG-4 standard, comprising:
presenting a streaming interface library containing a predetermined set of streaming control functions processing audiovisual objects, each of the control functions having predefined function calls; and
processing calls to the streaming interface library to decode and process streamed audiovisual objects according to the function calls.
12. The method of claim 11, wherein the step of processing comprises the step of executing a client application invoking the function calls.
13. The method of claim 11, further comprising the step of receiving user input, the user input invoking selected function calls.
14. The method of claim 11, wherein the streaming interface library comprises a visual decoding interface to decode visual objects contained in audiovisual bitstreams.
15. The method of claim 11, wherein the streaming interface library comprises a generation function for generating and outputting audiovisual data streams.
16. The method of claim 11, wherein the streaming interface library comprises an editing function for editing and manipulating audiovisual bitstreams.
17. The method of claim 11, wherein the streaming interface library comprises an interpreting function for interpreting audiovisual bitstreams.
18. The method of claim 11, wherein the step of processing comprises the step of adapting the execution of the bitstream interface library according to varying system resources.
19. The method of claim 11, further comprising the step of providing a client application interface, the client application interface invoking additional bitstream functions presented by cooperating client applications.
20. The method of claim 11, further comprising the step of providing a multimedia browser module employing the bitstream interface library for user viewing.
US09/055,929 1997-04-07 1998-04-07 System and method for generation and interfacing of bitstreams representing MPEG-coded audiovisual objects Expired - Lifetime US6044397A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US09/055,929 US6044397A (en) 1997-04-07 1998-04-07 System and method for generation and interfacing of bitstreams representing MPEG-coded audiovisual objects

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US4280197P 1997-04-07 1997-04-07
US09/055,929 US6044397A (en) 1997-04-07 1998-04-07 System and method for generation and interfacing of bitstreams representing MPEG-coded audiovisual objects

Publications (1)

Publication Number Publication Date
US6044397A true US6044397A (en) 2000-03-28

Family

ID=21923826

Family Applications (1)

Application Number Title Priority Date Filing Date
US09/055,929 Expired - Lifetime US6044397A (en) 1997-04-07 1998-04-07 System and method for generation and interfacing of bitstreams representing MPEG-coded audiovisual objects

Country Status (6)

Country Link
US (1) US6044397A (en)
EP (1) EP0909509B1 (en)
JP (2) JP4726096B2 (en)
CA (1) CA2257566C (en)
DE (1) DE69837833T2 (en)
WO (1) WO1998046024A1 (en)

Cited By (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20010015972A1 (en) * 2000-02-21 2001-08-23 Shoichi Horiguchi Information distributing method, information distributing system, information distributing server, mobile communication network system and communication service providing method
US20010025297A1 (en) * 2000-03-14 2001-09-27 Kim Sung-Jin User request processing method and apparatus using upstream channel in interactive multimedia contents service
US6337710B1 (en) * 1997-06-12 2002-01-08 Lsi Logic Corporation Graphical video editing system for regeneration of bitstreams
DE10041310A1 (en) * 2000-08-23 2002-03-07 Deutsche Telekom Ag Platform-independent streaming of multimedia contents for IP-based networks involves decoding compressed multimedia contents with Java applet automatically started by web browser
US6360195B1 (en) 1997-04-25 2002-03-19 Hongtao Liao Television or radio control system development
US20020057372A1 (en) * 1998-11-13 2002-05-16 Philips Electronics North America Corporation Method and device for detecting an event in a program of a video and/or audio signal and for providing the program to a display upon detection of the event
US20020062482A1 (en) * 2000-11-22 2002-05-23 International Business Machines Corporation System and method for selectable semantic codec pairs for very low data-rate video transmission
US6400400B1 (en) * 1997-07-30 2002-06-04 Sarnoff Corporation Method and apparatus for automated testing of a video decoder
US20020152462A1 (en) * 2000-08-29 2002-10-17 Michael Hoch Method and apparatus for a frame work for structured overlay of real time graphics
US6493023B1 (en) * 1999-03-12 2002-12-10 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Method and apparatus for evaluating the visual quality of processed digital video sequences
US20030008681A1 (en) * 2001-06-15 2003-01-09 Deutsche Telekom Ag Terminal device and method for using different services offered via a telecommunications network
US20030016747A1 (en) * 2001-06-27 2003-01-23 International Business Machines Corporation Dynamic scene description emulation for playback of audio/visual streams on a scene description based playback system
US20030030727A1 (en) * 2001-08-10 2003-02-13 Simon Gibbs System and method for enhancing real-time data feeds
US20030030658A1 (en) * 2001-08-10 2003-02-13 Simon Gibbs System and method for mixed reality broadcast
US20030030734A1 (en) * 2001-08-10 2003-02-13 Simon Gibbs System and method for transitioning between real images and virtual images
US20030046691A1 (en) * 2001-08-31 2003-03-06 Toshiyuki Nakagawa Data processing apparatus and method
US6535530B1 (en) * 1997-07-23 2003-03-18 Matsushita Electric Industrial Co., Ltd. Apparatus and method for demultiplexing multiplexed data
US6584125B1 (en) * 1997-12-22 2003-06-24 Nec Corporation Coding/decoding apparatus, coding/decoding system and multiplexed bit stream
US20030126603A1 (en) * 2001-12-29 2003-07-03 Kim Joo Min Multimedia data searching and browsing system
US20030156108A1 (en) * 2002-02-20 2003-08-21 Anthony Vetro Consistent digital item adaptation
US6611262B1 (en) * 1997-09-22 2003-08-26 Sony Corporation Generation of a bit stream containing binary image/audio data that is multiplexed with a code defining an object in ascii format
US20030163477A1 (en) * 2002-02-25 2003-08-28 Visharam Mohammed Zubair Method and apparatus for supporting advanced coding formats in media files
US6636931B2 (en) * 1998-01-06 2003-10-21 Pragmatic Communications Systems, Inc. System and method for switching signals over twisted-pair wires
US6654931B1 (en) * 1998-01-27 2003-11-25 At&T Corp. Systems and methods for playing, browsing and interacting with MPEG-4 coded audio-visual objects
US20030232366A1 (en) * 2000-10-12 2003-12-18 Marical, L.L.C. Polyvalent cation-sensing receptor in Atlantic Salmon
US20040006575A1 (en) * 2002-04-29 2004-01-08 Visharam Mohammed Zubair Method and apparatus for supporting advanced coding formats in media files
US20040010802A1 (en) * 2002-04-29 2004-01-15 Visharam Mohammed Zubair Generic adaptation layer for JVT video
US20040094020A1 (en) * 2002-11-20 2004-05-20 Nokia Corporation Method and system for streaming human voice and instrumental sounds
US6751655B1 (en) * 2000-04-18 2004-06-15 Sun Microsystems, Inc. Method and apparatus for transport of scenegraph information across a network
US20040139217A1 (en) * 2001-03-30 2004-07-15 Kidney Nancy G. One-to-one direct communication
US20040167925A1 (en) * 2003-02-21 2004-08-26 Visharam Mohammed Zubair Method and apparatus for supporting advanced coding formats in media files
US20050131930A1 (en) * 2003-12-02 2005-06-16 Samsung Electronics Co., Ltd. Method and system for generating input file using meta representation on compression of graphics data, and animation framework extension (AFX) coding method and apparatus
US20060209088A1 (en) * 2001-08-10 2006-09-21 Simon Gibbs System and method for data assisted chroma-keying
USRE39345E1 (en) * 1996-10-25 2006-10-17 Nec Corporation Audio/Video/Computer graphics synchronous reproducing/synthesizing system and method
US20070165139A1 (en) * 1997-02-14 2007-07-19 The Trustees Of Columbia University In The City Of New York Object-Based Audio-Visual Terminal And Bitstream Structure
US7461330B1 (en) * 1999-07-08 2008-12-02 At&T Intellectual Property Ii L.P. Methods and apparatus for integrating external applications into an MPEG-4 scene
US7613727B2 (en) 2002-02-25 2009-11-03 Sont Corporation Method and apparatus for supporting advanced coding formats in media files
US20100030852A1 (en) * 2007-02-02 2010-02-04 Streamezzo Method of Transmitting at Least One Content Representative of a Service, from a Server to a Terminal, and Associated Device and Computer Program Product
US8458597B1 (en) * 2010-02-04 2013-06-04 Adobe Systems Incorporated Systems and methods that facilitate the sharing of electronic assets
US9348829B2 (en) 2002-03-29 2016-05-24 Sony Corporation Media management system and process

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1999021337A2 (en) * 1997-10-17 1999-04-29 Koninklijke Philips Electronics N.V. Method of encapsulation of data into transport packets of constant size
US6351267B1 (en) 1998-12-10 2002-02-26 Gizmoz Ltd Fast transmission of graphic objects
US7240120B2 (en) * 2001-08-13 2007-07-03 Texas Instruments Incorporated Universal decoder for use in a network media player
FI20011871A (en) * 2001-09-24 2003-03-25 Nokia Corp Processing of multimedia data
EP1328127A1 (en) * 2002-01-09 2003-07-16 Beta Research GmbH Simultaneous feeding and synchronisation of audio and video data
KR100929073B1 (en) * 2005-10-14 2009-11-30 삼성전자주식회사 Apparatus and method for receiving multiple streams in portable broadcasting system

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5563648A (en) * 1994-04-28 1996-10-08 Thomson Consumer Electronics, Inc. Method for controlling execution of an audio video interactive program
US5794250A (en) * 1995-10-20 1998-08-11 Ncr Corporation Method and apparatus for extending existing database management system for new data types

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE69531265T2 (en) * 1994-03-19 2004-06-03 Sony Corp. Optical disc and method and apparatus for recording on and then reproducing information from that disc
JP3078215B2 (en) * 1995-01-06 2000-08-21 ミツビシ・エレクトリック・インフォメイション・テクノロジー・センター・アメリカ・インコーポレイテッド Display device

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5563648A (en) * 1994-04-28 1996-10-08 Thomson Consumer Electronics, Inc. Method for controlling execution of an audio video interactive program
US5794250A (en) * 1995-10-20 1998-08-11 Ncr Corporation Method and apparatus for extending existing database management system for new data types

Non-Patent Citations (16)

* Cited by examiner, † Cited by third party
Title
A. Basso et al., "Improved Data Access and Streaming Modes for the MPEG-4 File Format", pp. 1-12.
A. Basso et al., "MPEG-4 Integrated Intermedia Format (IIF): Basic Specification", International Organization for Standardization, ISO/IEC JTC1/SC29/WG11 MPEG98/M2978, Jan. 16, 1998, pp. 1-22.
A. Basso et al., Improved Data Access and Streaming Modes for the MPEG 4 File Format , pp. 1 12. *
A. Basso et al., MPEG 4 Integrated Intermedia Format (IIF): Basic Specification , International Organization for Standardization , ISO/IEC JTC1/SC29/WG11 MPEG98/M2978, Jan. 16, 1998, pp. 1 22. *
A. Eleftheriadis et al., "Stored File Format for MPEG-4", International Organization for Standardization, ISO/IEC JTC1/SC29/WG11 MPEG97/M2062, Apr. 1, 1997, pp. 1-4.
A. Eleftheriadis et al., "Stored File Format for Object-based Audio Visual Representation", pp. 1-8.
A. Eleftheriadis et al., Stored File Format for MPEG 4 , International Organization for Standardization , ISO/IEC JTC1/SC29/WG11 MPEG97/M2062, Apr. 1, 1997, pp. 1 4. *
A. Eleftheriadis et al., Stored File Format for Object based Audio Visual Representation , pp. 1 8. *
A. Griffin, "Video on the Net", Popular Mechanics, Mar. 1998, pp. 51-53.
A. Griffin, Video on the Net , Popular Mechanics , Mar. 1998, pp. 51 53. *
A. Puri et al., "APIs for MPEG-4 Systems", International Organization for Standardization, ISO/IEC JTC1/SC29/WG11 MPEG97/2133, Apr. 1997.
A. Puri et al., APIs for MPEG 4 Systems , International Organization for Standardization , ISO/IEC JTC1/SC29/WG11 MPEG97/2133, Apr. 1997. *
J. Heid, "Watch This: Streaming Video on Your Web Site", create WEB, Apr. 1998, pp. 109-112.
J. Heid, Watch This: Streaming Video on Your Web Site , create WEB , Apr. 1998, pp. 109 112. *
J. Laier et al., "Content-Based Multimedia Data Access in Internet Video Communication", First International Workshop on Wireless Image/Video Communications, Sep. 1996, pp. 126-133.
J. Laier et al., Content Based Multimedia Data Access in Internet Video Communication , First International Workshop on Wireless Image/Video Communications , Sep. 1996, pp. 126 133. *

Cited By (71)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
USRE39345E1 (en) * 1996-10-25 2006-10-17 Nec Corporation Audio/Video/Computer graphics synchronous reproducing/synthesizing system and method
US8854552B2 (en) 1997-02-14 2014-10-07 The Trustees Of Columbia University In The City Of New York Object-based audio-visual terminal and bitstream structure
US8922717B2 (en) 1997-02-14 2014-12-30 The Trustees Of Columbia University In The City Of New York Object-based audio-visual terminal and bitstream structure
US20070165139A1 (en) * 1997-02-14 2007-07-19 The Trustees Of Columbia University In The City Of New York Object-Based Audio-Visual Terminal And Bitstream Structure
US8421923B2 (en) 1997-02-14 2013-04-16 The Trustees Of Columbia University In The City Of New York Object-based audio-visual terminal and bitstream structure
US8115873B2 (en) 1997-02-14 2012-02-14 The Trustees Of Columbia University In The City Of New York Object-based audio-visual terminal and bitstream structure
US8917357B2 (en) 1997-02-14 2014-12-23 The Trustees Of Columbia University In The City Of New York Object-based audio-visual terminal and bitstream structure
US6360195B1 (en) 1997-04-25 2002-03-19 Hongtao Liao Television or radio control system development
US6337710B1 (en) * 1997-06-12 2002-01-08 Lsi Logic Corporation Graphical video editing system for regeneration of bitstreams
US6535530B1 (en) * 1997-07-23 2003-03-18 Matsushita Electric Industrial Co., Ltd. Apparatus and method for demultiplexing multiplexed data
US6400400B1 (en) * 1997-07-30 2002-06-04 Sarnoff Corporation Method and apparatus for automated testing of a video decoder
US6611262B1 (en) * 1997-09-22 2003-08-26 Sony Corporation Generation of a bit stream containing binary image/audio data that is multiplexed with a code defining an object in ascii format
US6584125B1 (en) * 1997-12-22 2003-06-24 Nec Corporation Coding/decoding apparatus, coding/decoding system and multiplexed bit stream
US6636931B2 (en) * 1998-01-06 2003-10-21 Pragmatic Communications Systems, Inc. System and method for switching signals over twisted-pair wires
US20040049791A1 (en) * 1998-01-06 2004-03-11 Shah Prasanna M. System and method for switching signals over twisted-pair wires
US6654931B1 (en) * 1998-01-27 2003-11-25 At&T Corp. Systems and methods for playing, browsing and interacting with MPEG-4 coded audio-visual objects
US8276056B1 (en) 1998-01-27 2012-09-25 At&T Intellectual Property Ii, L.P. Systems and methods for playing, browsing and interacting with MPEG-4 coded audio-visual objects
US9641897B2 (en) 1998-01-27 2017-05-02 At&T Intellectual Property Ii, L.P. Systems and methods for playing, browsing and interacting with MPEG-4 coded audio-visual objects
US7281200B2 (en) 1998-01-27 2007-10-09 At&T Corp. Systems and methods for playing, browsing and interacting with MPEG-4 coded audio-visual objects
US20020057372A1 (en) * 1998-11-13 2002-05-16 Philips Electronics North America Corporation Method and device for detecting an event in a program of a video and/or audio signal and for providing the program to a display upon detection of the event
US7124365B2 (en) * 1998-11-13 2006-10-17 Koninklijke Philips Electronics N.V. Method and device for detecting an event in a program of a video and/or audio signal and for providing the program to a display upon detection of the event
US6493023B1 (en) * 1999-03-12 2002-12-10 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Method and apparatus for evaluating the visual quality of processed digital video sequences
US9473770B2 (en) 1999-07-08 2016-10-18 At&T Intellectual Property Ii, L.P. Methods and apparatus for integrating external applications into an MPEG-4 scene
US10334278B2 (en) 1999-07-08 2019-06-25 At&T Intellectual Property Ii, L.P. Methods and apparatus for integrating external applications into an MPEG-4 scene
US8566692B2 (en) 1999-07-08 2013-10-22 At&T Intellectual Property Ii, L.P. Methods and apparatus for integrating external applications into an MPEG-4 scene
US20090092190A1 (en) * 1999-07-08 2009-04-09 At&T Corp. Methods and apparatus for integrating external applications into an mpeg-4 scene
US7461330B1 (en) * 1999-07-08 2008-12-02 At&T Intellectual Property Ii L.P. Methods and apparatus for integrating external applications into an MPEG-4 scene
US20010015972A1 (en) * 2000-02-21 2001-08-23 Shoichi Horiguchi Information distributing method, information distributing system, information distributing server, mobile communication network system and communication service providing method
US7346042B2 (en) * 2000-02-21 2008-03-18 Ntt Docomo, Inc. Information distributing method, information distributing system, information distributing server, mobile communication network system and communication service providing method
US7302464B2 (en) * 2000-03-14 2007-11-27 Samsung Electronics Co., Ltd. User request processing method and apparatus using upstream channel in interactive multimedia contents service
US20010025297A1 (en) * 2000-03-14 2001-09-27 Kim Sung-Jin User request processing method and apparatus using upstream channel in interactive multimedia contents service
US6751655B1 (en) * 2000-04-18 2004-06-15 Sun Microsystems, Inc. Method and apparatus for transport of scenegraph information across a network
DE10041310B4 (en) * 2000-08-23 2009-05-20 Deutsche Telekom Ag Method for platform-independent streaming of multimedia content for IP-based networks
DE10041310A1 (en) * 2000-08-23 2002-03-07 Deutsche Telekom Ag Platform-independent streaming of multimedia contents for IP-based networks involves decoding compressed multimedia contents with Java applet automatically started by web browser
US20020152462A1 (en) * 2000-08-29 2002-10-17 Michael Hoch Method and apparatus for a frame work for structured overlay of real time graphics
US20030232366A1 (en) * 2000-10-12 2003-12-18 Marical, L.L.C. Polyvalent cation-sensing receptor in Atlantic Salmon
US20020062482A1 (en) * 2000-11-22 2002-05-23 International Business Machines Corporation System and method for selectable semantic codec pairs for very low data-rate video transmission
US7219364B2 (en) * 2000-11-22 2007-05-15 International Business Machines Corporation System and method for selectable semantic codec pairs for very low data-rate video transmission
US7091997B2 (en) 2001-03-30 2006-08-15 Koninklijke Philips Electronics N.V. One-to-one direct communication
US20040139217A1 (en) * 2001-03-30 2004-07-15 Kidney Nancy G. One-to-one direct communication
US7068294B2 (en) 2001-03-30 2006-06-27 Koninklijke Philips Electronics N.V. One-to-one direct communication
US7793327B2 (en) * 2001-06-15 2010-09-07 Deutsche Telekom Ag Terminal device and method for using different services offered via a telecommunications network
US20030008681A1 (en) * 2001-06-15 2003-01-09 Deutsche Telekom Ag Terminal device and method for using different services offered via a telecommunications network
US7216288B2 (en) * 2001-06-27 2007-05-08 International Business Machines Corporation Dynamic scene description emulation for playback of audio/visual streams on a scene description based playback system
US20030016747A1 (en) * 2001-06-27 2003-01-23 International Business Machines Corporation Dynamic scene description emulation for playback of audio/visual streams on a scene description based playback system
US7339609B2 (en) 2001-08-10 2008-03-04 Sony Corporation System and method for enhancing real-time data feeds
US20030030727A1 (en) * 2001-08-10 2003-02-13 Simon Gibbs System and method for enhancing real-time data feeds
US20060209088A1 (en) * 2001-08-10 2006-09-21 Simon Gibbs System and method for data assisted chroma-keying
US20030030658A1 (en) * 2001-08-10 2003-02-13 Simon Gibbs System and method for mixed reality broadcast
US20030030734A1 (en) * 2001-08-10 2003-02-13 Simon Gibbs System and method for transitioning between real images and virtual images
US7173672B2 (en) 2001-08-10 2007-02-06 Sony Corporation System and method for transitioning between real images and virtual images
US8457350B2 (en) 2001-08-10 2013-06-04 Sony Corporation System and method for data assisted chrom-keying
US8022965B2 (en) 2001-08-10 2011-09-20 Sony Corporation System and method for data assisted chroma-keying
US20030046691A1 (en) * 2001-08-31 2003-03-06 Toshiyuki Nakagawa Data processing apparatus and method
US7690018B2 (en) * 2001-08-31 2010-03-30 Canon Kabushiki Kaisha Data processing apparatus and method
US20030126603A1 (en) * 2001-12-29 2003-07-03 Kim Joo Min Multimedia data searching and browsing system
US20030156108A1 (en) * 2002-02-20 2003-08-21 Anthony Vetro Consistent digital item adaptation
US20030163477A1 (en) * 2002-02-25 2003-08-28 Visharam Mohammed Zubair Method and apparatus for supporting advanced coding formats in media files
US7613727B2 (en) 2002-02-25 2009-11-03 Sont Corporation Method and apparatus for supporting advanced coding formats in media files
US9348829B2 (en) 2002-03-29 2016-05-24 Sony Corporation Media management system and process
US20040010802A1 (en) * 2002-04-29 2004-01-15 Visharam Mohammed Zubair Generic adaptation layer for JVT video
US20040006575A1 (en) * 2002-04-29 2004-01-08 Visharam Mohammed Zubair Method and apparatus for supporting advanced coding formats in media files
US7831990B2 (en) 2002-04-29 2010-11-09 Sony Corporation Generic adaptation layer for JVT video
US20040094020A1 (en) * 2002-11-20 2004-05-20 Nokia Corporation Method and system for streaming human voice and instrumental sounds
US20040167925A1 (en) * 2003-02-21 2004-08-26 Visharam Mohammed Zubair Method and apparatus for supporting advanced coding formats in media files
US20040199565A1 (en) * 2003-02-21 2004-10-07 Visharam Mohammed Zubair Method and apparatus for supporting advanced coding formats in media files
US7561745B2 (en) * 2003-12-02 2009-07-14 Samsung Electronics Co., Ltd. Method and system for generating input file using meta representation on compression of graphics data, and animation framework extension (AFX) coding method and apparatus
US20050131930A1 (en) * 2003-12-02 2005-06-16 Samsung Electronics Co., Ltd. Method and system for generating input file using meta representation on compression of graphics data, and animation framework extension (AFX) coding method and apparatus
US20100030852A1 (en) * 2007-02-02 2010-02-04 Streamezzo Method of Transmitting at Least One Content Representative of a Service, from a Server to a Terminal, and Associated Device and Computer Program Product
US9560401B2 (en) * 2007-02-02 2017-01-31 Streamezzo Method of transmitting at least one content representative of a service, from a server to a terminal, and associated device and computer program product
US8458597B1 (en) * 2010-02-04 2013-06-04 Adobe Systems Incorporated Systems and methods that facilitate the sharing of electronic assets

Also Published As

Publication number Publication date
JP5084644B2 (en) 2012-11-28
EP0909509A4 (en) 2004-09-22
DE69837833T2 (en) 2008-01-31
WO1998046024A1 (en) 1998-10-15
CA2257566A1 (en) 1998-10-15
JP4726096B2 (en) 2011-07-20
JP2000513178A (en) 2000-10-03
CA2257566C (en) 2002-01-01
EP0909509A1 (en) 1999-04-21
DE69837833D1 (en) 2007-07-12
JP2008289187A (en) 2008-11-27
EP0909509B1 (en) 2007-05-30

Similar Documents

Publication Publication Date Title
US6044397A (en) System and method for generation and interfacing of bitstreams representing MPEG-coded audiovisual objects
JP4959504B2 (en) System and method for interfacing MPEG coded audio-visual objects capable of adaptive control
US10334278B2 (en) Methods and apparatus for integrating external applications into an MPEG-4 scene
US7529806B1 (en) Partitioning of MP3 content file for emulating streaming
EP1131930B1 (en) Partitioning of file for emulating streaming
Signes et al. MPEG-4's binary format for scene description
US7634727B2 (en) System for abstracting audio-video codecs
US20030046691A1 (en) Data processing apparatus and method
CN113615205A (en) Method, apparatus and computer program for signaling available portions of packaged media content
US7307553B2 (en) MPEG-4 encoding/decoding method, medium, and system
Black et al. A compendium of robust data structures
EP1912438A2 (en) System and method for interfacing MPEG-coded audiovisual objects permitting adaptive control
JP2005176094A (en) Data processor, data processing method, program and storage medium
Herpel et al. Olivier Avaro Deutsche Telekom-Berkom GmbH, Darmstadt, Germany Alexandros Eleftheriadis Columbia University, New York, New York
JP2005159878A (en) Data processor and data processing method, program and storage medium
Joung et al. XMT tools for interactive broadcasting contents description
KR20040016566A (en) Method for representing group metadata of mpeg multi-media contents and apparatus for producing mpeg multi-media contents
Kalva Object-Based Audio-Visual Services
KR20090110200A (en) Method and apparatus for providing/receiving user interface

Legal Events

Date Code Title Description
AS Assignment

Owner name: AT&T CORP., NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PURI, ATUL;SCHMIDT, ROBERT LEWIS;REEL/FRAME:009321/0877;SIGNING DATES FROM 19980605 TO 19980610

AS Assignment

Owner name: COLUMBIA UNIVERSITY, NEW YORK

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ELEFTHERIADIS, ALEXANDROS;FANG, YIHAN;REEL/FRAME:010642/0447;SIGNING DATES FROM 20000210 TO 20000211

STCF Information on status: patent grant

Free format text: PATENTED CASE

FEPP Fee payment procedure

Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

FPAY Fee payment

Year of fee payment: 12