WO2000073914A1 - Synchronized spatial-temporal browsing of images for selection of indexed temporal multimedia titles - Google Patents

Synchronized spatial-temporal browsing of images for selection of indexed temporal multimedia titles Download PDF

Info

Publication number
WO2000073914A1
WO2000073914A1 PCT/US2000/013561 US0013561W WO0073914A1 WO 2000073914 A1 WO2000073914 A1 WO 2000073914A1 US 0013561 W US0013561 W US 0013561W WO 0073914 A1 WO0073914 A1 WO 0073914A1
Authority
WO
WIPO (PCT)
Prior art keywords
multimedia
presentation
multimedia data
data items
presented
Prior art date
Application number
PCT/US2000/013561
Other languages
French (fr)
Inventor
John Raymond Nicol
Christopher Michael Martin
James Edward Paschetto
Kent Barrows Wittenburg
Original Assignee
Gte Laboratories Incorporated
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/560,006 external-priority patent/US7281199B1/en
Application filed by Gte Laboratories Incorporated filed Critical Gte Laboratories Incorporated
Priority to AU48553/00A priority Critical patent/AU4855300A/en
Publication of WO2000073914A1 publication Critical patent/WO2000073914A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/44Browsing; Visualisation therefor
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • G06F16/438Presentation of query results
    • G06F16/4387Presentation of query results by the use of playlists
    • G06F16/4393Multimedia presentations, e.g. slide shows, multimedia albums
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/48Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually

Definitions

  • This application generally relates to computer systems, and more particularly to computer executed techniques for locating and selecting information.
  • Computer systems may be used in a variety of applications and environments. For example, in commercial applications of computer systems, a computer may assist a user in making a decision or choice based on information that is viewed using the computer. A computer storing information may be used to display information to a user. Subsequently, the user may make a decision or selection in light of the information presented.
  • One type of information from which a user may make a selection is multimedia data.
  • video key framse may serve as reference framse and be used to index into videos.
  • a problem may be encountered when the key frames are similar in that a user is associating a key frame with a distinct portion of a video segment.
  • it may not be clear which one of the media streams to use as an index. In this instance, for example, use of video key frames as an index into portions of multimedia data for selecting and browsing may not be adequate.
  • Browsing and selecting techniques such as those employed by an Internet browser like Internet Explorer or Netscape Navigator, may be used in conjunction with static data or content information. However, these techniques may not be particularly suitable for use with temporal based multimedia information such as the foregoing video information and other types of mulitmedia data that may be used as indices into multimedia presentations being browsed.
  • Figure 1 is an example of an embodiment of a system
  • Figure 2 is an example of a block diagram of software components included in the system of Figure 1 ;
  • Figure 2 A is an example of one embodiment of the hierarchical data file of Figure 2;
  • Figure 3 is an example of an embodiment of a block diagram illustrating the flow of control in producing and using various software components included in the system of Figure 1 ;
  • Figure 4 is a flowchart of an example embodiment of method steps performed at run time when a client browser 16 issues a request for the software browsing tool;
  • Figure 5 is an example of an embodiment of the browsing software tool that may be included in the system of Figure 1 ;
  • Figure 6 is an example of a user interface display for a generated browser software tool;
  • Figure 7 is an example of an embodiment of a user interface displayed with a detailed outline
  • Figure 8 is an example of an embodiment of a user interface of the software browsing tool using an auto-scrolling method for presenting multimedia data
  • Figure 9 is an example of an embodiment of a user interface of a software browsing tool that presents multimedia data as a stacked slide show
  • Figure 10 is an example of an embodiment of a user interface displayed with the browser tool using the flyby effect
  • Figure 11 is an example of an embodiment of a layout technique that may be used with the flyby effect in presenting a group of multimedia data
  • Figure 12A is an example of an embodiment of a top-down view of multimedia data presented using the flyby effect and layout technique of Figure 1 1
  • Figure 12B is an example of an embodiment of a layout technique that may be used in a multimedia data presentation
  • Figure 13 is an example of an embodiment of a schematic for a technique for presenting multimedia data as in Figure 12B
  • Figure 14 is an example of an embodiment of user controls as included in various user interfaces for the browsing tool
  • Figure 16 is a flowchart of method steps of one embodiment for generating and using indices with a synchronized multimedia presentation
  • Figure 17 is an example of an embodiment of a database schema record for indices created and used in connection with multimedia presentation to be browsed using the browsing tool
  • Figure 18 is a more detailed example of the database schema record of Figure 17;
  • Figure 19 is a flowchart of steps of a method for selecting a multimedia presentation using the browsing tool and multimedia presentation indices from the database;
  • Figure 20 is an example of an embodiment of a data structure used to associate categories with one or more multimedia presentation items;
  • Figure 21 is an example of an embodiment of a data structure used to associate indices with each multimedia presentation item
  • Figure 22 is an example of an embodiment of a user interface display of video titles using the browsing tool
  • Figure 23 is an example of an example of an embodiment of a screen displayed when a selection of a movie title has been made using the browsing tool
  • Figure 24 is an example of an embodiment of a user interface display of a multimedia presentation using the browsing tool
  • Figure 25 is an example of an embodiment of a screen displayed when a selection of a multimedia presentation has been made using the browsing tool
  • Figure 26 is an example of an embodiment of a user interface display of multimedia shopping titles displayed using the browsing tool
  • Figure 27 is an example of an embodiment of a screen displayed when a selection of a multimedia shopping title has been made using the browsing tool
  • Figure 28 is an example of an embodiment of a system that includes a journaling application for recording and playing back a multimedia presentation
  • Figure 29 is an example of an embodiment of a user interface display of a journaling application
  • Figure 30 is an example of
  • Figure 31 is an example of an embodiment of a time line as may be used to represent the time coordination of events in a multimedia presentation; and
  • Figure 32 is a flowchart of method steps of an embodiment for synchronizing interrelated components of a multimedia presentation.
  • a subset of multimedia presentations are determined in accordance with predetermined criteria.
  • One or more multimedia data items are determined.
  • Each of the one or more multimedia data items is an index into a corresponding one of the subset of multimedia presentations.
  • the one or more multimedia data items are viewed using a browser.
  • Direction and speed of the one or more multimedia data items viewed is controlled.
  • a first of the one or more multimedia data items is selected. Control is transferred to machine executable code associated with a first multimedia presentation having the first multimedia data item as an index.
  • a method and apparatus for selecting a multimedia presentation is a method and apparatus for selecting a multimedia presentation.
  • One or more indices are determined for a multimedia presentation.
  • the multimedia presentation has a first media stream and a second media stream. A portion of the first media stream is used as an index into a portion of said second media stream.
  • Information about the indices is stored in a database.
  • Multimedia objects are displayed in accordance with the information about the indices in the database using a browser. Direction and speed of the multimedia objects displayed are controlled. A first of the multimedia objects is selected.
  • a software program is invoked for presenting the multimedia presentation having the first multimedia object as an index.
  • a subset of multimedia presentation is determined in accordance with predetermined criteria.
  • One or more multimedia data items is determined.
  • Each of the one or more multimedia data items is an index into a corresponding one of the subset of multimedia presentations.
  • One or more multimedia data items is viewed using a browser to select a multimedia presentation.
  • Direction and speed of the one or more multimedia data items viewed is controlled.
  • a first of the one or more multimedia data items associated with the multimedia presentation is selected.
  • a computer system 10 is shown to include a network 12 by which computer nodes 14- 18 may communicate with each other.
  • one of the computer nodes such as node 15 acts as a client node upon which a user executes a client browser 16.
  • the client node 15 communicates with one or more other nodes in the system 10 in which the one or more other nodes in the computer system 10, such as node 14, may act as a server node providing certain functions or responding to certain requests made by the client browser executing in the client node 15.
  • system of Figure 1 may include one or more computer nodes and the nodes, if there are more than one, may communicate through any type of communications medium, such as a network, a hardwire connection, or other means of communication known to those skilled in the art.
  • each of the computer nodes in the system 10 may be any type of computer processor, such as a commercially available personal computer, or a larger scale multi-user computer system.
  • the software components in Figure 2 may be used to produce the files needed to build a software browsing tool. Additionally, some of the components included in Figure 2 are files that may be used by the software browsing tool once complete.
  • the software browsing tool that will be described in paragraphs that follow is a tool providing overviews of dynamically specified data for the purpose of, for example, previewing or assessing data.
  • the tool presents an organizational structure by which a user may view various data components and objects, such as those associated with image -based and/or multimedia data presentations.
  • the hierarchical data file 20 generally includes a description of the various data components to be presented and viewed by a user in some type of hierarchical organization.
  • the hierarchical data file may include a list of the various URLs specifying images and/or other multimedia data files to be presented to a user. Additionally, various URLs are grouped together to represent the grouping hierarchy by which these components are to be presented.
  • the hierarchical data file may include a nesting structure nesting various groups of one or more of the URL identified files presented to the user.
  • the user interface components 22 generally include software used to produce the user interface displays with the browsing software tool.
  • user interface components may include, for example, JavaScript code and data, and dynamic HTML files.
  • Multimedia data files 24 generally include multimedia data to be presented to the user in conjunction with the user interface.
  • the multimedia data files include that data which is organized and represented by the hierarchical data file.
  • other types of data files such as multimedia data files including audio files may also be included in other embodiments.
  • the translation tool 26 is generally used to produce other pieces of software which are used in displaying the user interface and associated data files.
  • the translation tool 26 includes Java code to produce files which will be described in conjunction with Figure 3.
  • the hierarchical data file includes data 150 that may be grouped together as previously set forth in conjunction with Figure 2.
  • the hierarchical data file for each grouping of one or more multimedia files, includes NAME/TITLE INFORMATION, one or more Related
  • the NAME/TITLE INFORMATION is a text description of the multimedia files that are the Presentation URLs.
  • the NAME/TITLE INFORMATION may be included, as in a menu area, as will be described in following paragraphs.
  • the Presentation URLs are one or more multimedia files presented to the user in a presentation area, for example, of a user interface.
  • the one or more Related URLs may annotate or further describe the one or more Presentation URLs.
  • the Presentation URLs may identify images of various types of cars by a manufacturer.
  • One of the Related URLs may identify a website of the manufacturer.
  • the NAME/TITLE INFORMATION may be a text description identifying the manufacturer and type of cars presented.
  • FIG. 3 shown is an example of an embodiment of the various data files used and produced by the translation tool 32 to display the user interface and provide the software browsing tool.
  • the hierarchical data file 30 is used as input to the translation tool 32.
  • the translation tool 32 may produce a variety of software output files.
  • the software output files are included in the software browsing tool used to present data.
  • the actual output produced by the translation tool is in accordance with the actual implementation of the various components of the software browsing tool.
  • the translation tool produces the VRML code 34.
  • VRML code 34 is code written in a language called Virtual Reality Modeling Language.
  • a VRML player which interprets or executes the VRML code 34 may be included in browsers such as the Netscape 4.0 or the Internet
  • the VRML player for example, is one component required to execute the VRML code which implements one of the components of the browsing software which will be described in paragraphs that follow.
  • the translation tool 32 may also produce JavaScript code and data used with, for example, dynamic HTML files 36. Generally, if an implementation or embodiment includes the use of HTML files or VRML code 34, the translation tool 32 may generate JavaScript code and data.
  • the translation tool 32 may also produce Java Code 38.
  • Java Code 38 may be used as an alternative to writing various components of the software browsing tool using the dynamic HTML files.
  • FIG. 4 shown is a flowchart of an example of an embodiment of the method steps of what happens at run time when a client browser 16 issues a request for the software browsing tool.
  • the client makes a request.
  • the client browser 16 on node 15 may issue a request using network 12 to communicate to another server node, such as node 14.
  • the user interface components located on a node, such as server node 14 are loaded into and executed on the client system by the client browser 16.
  • execution of the user interface components causes the client browser 16 to make a request to the server node 14 for the hierarchical data file.
  • execution of the user interface components causes the translation tool to execute producing the various software components of the software browsing tool and subsequently sending these software components produced to the client node 15.
  • these translated files produced by the translation tool 32 are executed by the client browser on the client system 15 causing images and other multimedia files to be loaded in accordance with a particular embodiment.
  • various communication requests may need to be issued by the server node, for example, to obtain the necessary files to produce the software components.
  • the precise images and other multimedia files loaded are in accordance with each particular embodiment. For example, use of the
  • VRML player causes all images to be loaded by default.
  • Other software tools may have other defaults as well as variants which may allow image files and other multimedia data files to be loaded on demand. It should also be noted that in an embodiment where all of the files in processing occurs on the client system, no additional communication may be needed since the client and server node are the same.
  • the software browsing tool of Figure 5 includes driver commands and data 40 and various software components 46-50.
  • the software components may vary with each particular embodiment of the software browsing tool and the functions provided, as well as the tools used to implement the software browsing tool.
  • the driver program and data 40 includes a driver program or script 42 and current focus data 44.
  • the driver program or script 42 is a set of commands or statements which drive or control the software browsing tool.
  • the current focus data 44 is generally a common shared data area describing the focus of the data presented to the user.
  • the current focus of the data being presented may be changed as the user cycles through and views the various multimedia data items. As this happens, the focus changes as the user examines different images or other multimedia data.
  • the focus data 44 is updated by a signaling handler as specified in the driver program or script 42.
  • a mechanism exists in the driver program or script providing a control for updating the focus data 44 and additionally signaling all of the components as to the change in the current focus.
  • the various components then update their local copy of the focus data and redisplay their various portions of the user interface.
  • the driver/program script 42 is a JavaScript and the current focus data is represented as a JavaScript object.
  • the driver program or script 42 may be one of the JavaScript code and data components produced by the process of Figure 3 by the translation tool 32.
  • the VRML runtime component 46 may include VRML code 34 as produced by the translation tool 32 of Figure 3.
  • control bar component 48 and the sound player component 50 may be implemented using a variety of different coding languages.
  • Java code as may be produced by the translation tool 32, may be the language of implementation for these components.
  • one or more of these components 48-50 may be implemented using
  • each of the components 46-50 of Figure 5 for the browsing software tool are generally independent, pluggable pieces that may be removed or added independent of the other components.
  • each component maps to a particular component of the user interface displayed and is responsible, for example, for updating and performing other functions within that particular portion of the user interface.
  • the hierarchical data file 30 of Figure 3 may be produced manually or in an automated fashion, such as by a software tool that may be implemented using one of many different languages known to those skilled in the art.
  • various directory structures and subdirectories may be used to group files in accordance with the hierarchy description to be included in the hierarchical data file.
  • a software program written in a commercially written programming language for example, may use system functions to inquire as to the contents of the various directories and subdirectories and accordingly, produce a hierarchical data file in accordance with the directory and subdirectory structure.
  • Other embodiments may include different ways of producing the hierarchical data file in an automated fashion in accordance with a particular system of Figure 1.
  • Figure 6 shown in an example of an embodiment of a user interface as displayed by the software browsing tool.
  • Figure 6 shows an example of the browsing tool used in the application of the domain of Internet shopping.
  • the user interface 60 includes a menu or outline portion 62, a presentation area 64, and user control buttons 66.
  • the outline or menu portion 62 displays various topics or text descriptions which accompany the items, images, or other multimedia data being presented in the presentation area 64.
  • control display 66 includes buttons which provide for backward and forward review of the multimedia data presented in the presentation area 64. Accordingly, the active or highlighted portion, for example, of the outline or menu area 62 is synchronized with the presentation in accordance with the user positioning of the controls. If the user, for example, wants to preview multimedia data items that were already presented, the rewind button from the user control area 66 would be used also causing the menu portion 62 to have its active item updated in accordance with the multimedia data items presented in the presentation area 64.
  • the active or current focus is denoted by highlighting, for example, a particular item in the menu portion 62.
  • the current outline or menu item has an arrow 63 next to it in the upper left hand corner.
  • the arrow in the left hand corner advances accordingly in the menu portion 62.
  • Other techniques may be used to indicate an active or current focus on the outline area 62.
  • buttons 62 include highlighting a particular area on the menu, or showing a color contrast from the rest of the menu items.
  • the content of the multimedia data presented has to do with promotional advertisements generally associated with online or virtual storage such as those available through the Internet that had been categorized by product type as well as occasion.
  • promotional advertisements generally associated with online or virtual storage such as those available through the Internet that had been categorized by product type as well as occasion.
  • the system generates the browser tool for the set of stores that have a promotional offering in the category selected by the option 65.
  • images and/or other multimedia data representing the promotional content are presented one after the other in a semi-random spatial layout in the presentation area 64. Transitions between stores clear the presentation area.
  • the store whose promotional items are currently being presented in the presentation area 64 are indicated in the listing at the left by the appearance of a red arrow, as indicated in the outline or menu portion 62.
  • the user may control the speed of the presentation of the images and other multimedia data by choosing among the arrow controls 66.
  • the greater distance the cursor is placed from the center of the control area the greater the speed in a particular direction.
  • two sets of arrows in the control area 66 go to the right indicating forwarding of the multimedia data presentation
  • two sets of display arrows go to the left indicating reversing the multimedia data presentation.
  • the farther to the left the cursor is placed over the arrows the faster in reverse the multimedia data in the presentation area is displayed. This also parallels the fast forwarding controls as indicated by the arrows to the right in the control area 66.
  • multimedia data presented in the presentation area 64 of Figure 6 and in other user interfaces which will be described in paragraphs that follow may be presented in a variety of different presentation arrangements.
  • the multimedia data is presented in a stacking arrangement, as will be described in more detail below.
  • Figure 7 shown is a user interface of the browser software tool with a variant including more detail in the text outline portion 72.
  • Figure 7 includes a presentation area 76, user control 74, and an outline or menu portion 72. These areas are similar to those as previously described in conjunction with Figure 6 user interface 60.
  • This variant of the browser is also for the shopping domain in which the listing of the stores in the left hand portion includes store titles as well as the names of the products being promoted.
  • the user interface 60 included only stores in the outline or menu area.
  • the outline or menu area 72 includes a store name, such as DisneyTM 71 , as well as the product being promoted such as Thanksgiving
  • PoohTM73 In this variant, products are shown one at a time in order to make synchronization with the presentaton area 76 and the menu item 72 straight forward with the detailed hierarchical text outline listing.
  • the browser variant 70 of Figure 7 may present one image or other multimedia data element at a time. This is in contrast to the user interface 60 of Figure 6, in which multiple multimedia data objects or elements are presented which may overlap or be superimposed upon one another.
  • multimedia data presented in the presentation areas 76, 64, and other user interfaces may be "dragged and dropped" for performing other user functions.
  • the user may drag and drop an item from the presentation area to place in a virtual shopping basket for items which the user may wish to purchase at the end of their viewing or browsing session.
  • a user may be connected, as by a hyper link, to a different Internet website. Selection may also have other meanings depending on the application of the viewing or browsing session.
  • selecting a particular item in the presentation area may mean to select a movie and view a portion of a movie trailer, for example. If it were a TV channel selection, an image selection from the television show on each particular channel may be presented in the presentation area 76. The user may select a TV channel and the particular TV program for current viewing.
  • multimedia files may also be executed, for example, while an image is being viewed.
  • a sound file in addition to an image file may be played to a user when viewing images, for example, of a particular movie.
  • a user may jump to a particular section in the sequence by selecting an item from the menu For example, if the user's focus or current point is at the top item as in Figure 6, a user may advance to the end of the menu item and accordingly cause presentation of the multimedia data associated with that item by selecting the last item from the menu area This is a shortcut method by which a user may quickly advance to a particular area in the menu to look at and view the images and other multimedia data associated with that particular item in the menu selection.
  • the user interface of Figure 6 may generally be described as a slide show presentation by which the user may begin a multimedia presentation associated with the first item in the menu area 62.
  • An embodiment of this may automatically advance at a set speed from the beginning to the end of the menu and accordingly present multimedia data in the presentation area 64 at a given pace.
  • the user may control the presentation of the multimedia data by using the control area 66 This is similar to the way in which a movie, for example, may be viewed in which the play button causes the images of the movie to be presented at a particular speed.
  • the multimedia presentation of Figure 7 generally includes a hierarchical level of menus displayed in the area 72
  • there is a nesting of several levels of menu items Generally, the outermost level menu is denoted as level 1 and the level numbers increase sequentially to 3 and the like as the nesting gets deeper In this particular display 72, while an innermost level is displayed, the outermost levels also stay constant on the menu
  • the nesting level was level 3, and the menu items spanned more than one screenful, while the additional items from the level 3 menu were displayed, the level 1 and level 2 items would appear constant on the screen for the user so that the user is provided with information indicating where the current focus is for the menu items being displayed
  • the user is provided with context feedback information describing where at any particular time multimedia data m the presentation area is located relative to the menu items 72.
  • the presentation area 76 of Figure 7 may present images and other multimedia data using a variety of techniques.
  • Multimedia data may be presented in a temporal fashion, one multimedia data item at a time with subsequent data items replacing or overlaying previous multimedia data items, in synchronization with a sound file and the more detailed menu 72, for example.
  • Multimedia data may also be presented in a rolling or scrolling fashion in an upward or downward direction. This technique is further described below.
  • the user interface 75 includes user control 77, a presentation area 75, with multimedia data presented, such as multimedia data items 202 and 204, and a menu or outline area 78.
  • multimedia data presented such as multimedia data items 202 and 204
  • a menu or outline area 78 the user scrolls through a series of multimedia data items which are predetermined in a format layout out such as in a top to bottom fashion.
  • the motion of controlling or scrolling through the multimedia data using control 77 scrolls along in the x-y plane as indicated by the arrow on the right hand side for forward and reverse direction of presenting the multimedia data items.
  • a portion of multimedia data items 202 and 204 are shown as extending beyond the user interface screen 75 for the purposes of demonstration only.
  • a portion of the multimedia data 202 would be presented which is off the screen for the current point of view.
  • a portion of the multimedia data 204 which is shown as off the screen currently would be included on the presentation area 52 displacing a portion of the other multimedia data items, such as 202.
  • the user has a point of view or focus which is the size of the presentation area 52 which scrolls downward through a prearranged presentation of multimedia data in a layout fashion.
  • a user may view portions of the multimedia data in the presentation area 52. As described before in conjunction with other figures, the user may advance to a different position on the scroll by selecting a particular menu item. This is in addition to using the user control 77 to rewind and fast forward to the multimedia data contained in the scroll.
  • Level- 1 and “Level-2” of menu area 78 may be “active” in that they may be selected and used to modify the current focus or point of view of the user with regard to the data presented in the presentation area 52.
  • FIG. 9 shown is an example of an embodiment of the user interface for the browser tool which presents multimedia data in the presentation area 52 in a stacked slide show fashion.
  • multimedia data items which may be presented in the presentation area 52 are slightly stacked upon one another for viewing at a predetermined time interval.
  • the user may control with the user controls 77 how fast or slow the multimedia data items are stacked upon one another in the presentation area 52.
  • the multimedia data may also appear in an unstacked fashion in the presentation area 52 such that the multimedia data items do not overlap each other.
  • the scrolling effect had the images and/or other multimedia data presented in a predetermined top to bottom fashion with the user interface presentation area acting as a window clipping or viewing a portion of this predetermined area.
  • the stacked slide show is similar to the appearance of having a deck of cards where each multimedia data item corresponds to a card in the deck in which the cards are placed one on top of each other in the presentation area. Fast forwarding or rewinding through a stack slide show of multimedia data results in cards being removed or placed on top of the deck. Similarly, it would result in multimedia data items being stacked or layed on top of one another or removed from the stack.
  • FIG. 10 shown is an example of a user interface of the browser variant in which the presentation area or the display area incorporates a three-dimensional fly through or flyby effect as opposed to the previously described scrolling effect in various slide show effects.
  • the user may control the presentation in the presentation area 86 using the control arrows 84
  • the user may select various categories of multimedia data items to be presented using the buttons indicated in the selection area 82 of the user interface display 80
  • a user may select a particular multimedia data item from the presentation area 86 which causes the user to jump to the page where the multimedia data item is from
  • selecting an item from the presentation area such as with the mouse, a user may use the hyperlink Internet connection by which they are connected to the page on the Internet upon which an image or other multimedia data item may be located
  • the position of the user's mouse cursor determines the speed of the presentation which essentially has an infinite range from 0 to the maximum that the display device is capable of achieving Audio feedback of the
  • FIG. 1 1 shown is an example of an embodiment of a frontal point of view of multimedia data presented using the flyby effect of Figure 10
  • the presentation arrangement 182 is one technique that may be used to present images in the presentation area 86 of Figure 10
  • Data items 160-170 are presented in as a group of data Data items are presented one at a time in a counter-clockwise direction in a circular arrangement
  • the center of each data item 160-170 is placed on the circumference of the circle 184 as it appears in the presentation area 86
  • circle 184 does not actually appear in the presentation area but is displayed in Figure 1 1 to indicate placement of the data items 160-170
  • Each of the data items 160-170 are placed equidistant apart such that each of the angles 172-
  • This technique may be generalized for any number of data items such that they are placed equidistant apart in a circular arrangement in a counterclockwise order.
  • Subsequent groups of data items are presented in an arrangement with a random angular shift in the counterclockwise direction.
  • the position of the first data item 160 is shifted by a random amount in the counterclockwise direction.
  • the placement of subsequently presented data items in the same group are positioned with the same angular shift.
  • each of the data item groups 122-128 represents a group of data items presented similar to those in the group 182 of Figure 1 1.
  • each of the groups 122-128 is a "slice" or page of data items presented using the circular layout technique described in conjunction with Figure 1 1.
  • the procedure for laying out multimedia data in a 3D flyby effect includes defining a certain area and placing the multimedia data for presentation within that defined area.
  • a user's view or perspective at a particular point in time such as a snapshot of the multimedia data in Figure 1 1 is depicted as having the user move along the arrow indicated in the left-hand side of Figure 12A on the x-y axis 69.
  • a user's point of view or snapshot of the multimedia data such as those presented in Figure 10 in the presentation area 86 represent a viewpoint of the multimedia taken from user perspective.
  • the data items 124, 126, and 128 will appear to fly past the user and the data item 122 will appear to come closer to the user.
  • the data items will appear to fly by the user in reverse order with a perspective or focal point into the screen.
  • the data items presented have a perspective such that the farther the multimedia data items are in the fast forward direction, the farther away they appear in the presentation to the user and appear to go to a focal point into the screen.
  • multimedia data items are viewed, they pass by the user on the presentation area 67 giving the user the impression that the multimedia data items are moving past where the user is out of the visual path.
  • the multimedia data items are presented or refreshed in reverse order. This gives the appearance as if the data items are coming from behind the user to the current focus on the screen and, as the rewinding continues, the multimedia data items move into the screen such as at a distant point.
  • the top or downward view of the user perspective in Figure 12A generally includes a defined area 67 where groups of multimedia data items 122-128 may be positioned.
  • the user's perspective or focus moves along the x-y axis, as indicated by the arrow 69.
  • the user's view point may be associated with a point along the x-y axis 69.
  • the multimedia data items are presented to the user from the user's perspective at a particular point along the x-y axis denoted by the arrow 69 as if the user were sitting in front of the defined area 67 with all of the multimedia data items coming towards the user or away from the user, as controlled by the control area 66 of Figure 6.
  • FIG. 12B shown is an example of an embodiment of another technique for presenting data in a presentation area 120.
  • the presentation area 120 may be included in one of the user interface displays such as 80.
  • FIG. 120 What will be described in conjunction with the next two figure is a technique for presenting multimedia data in a group layout in which there is a parent and child relationship between various multimedia data items to be presented on the presentation area 120.
  • there is a group of data items 190-198 with the parent 198 being located in the center of the presentation area and the children 190-196 being located surrounding the parent.
  • the parent is presented in the center of the group with children being placed in four quadrants beginning with the upper left relative to the center of the group which includes the parent node. It should generally be noted that if there were only three children for a particular group, the children would be placed in order beginning with the upper left quadrant proceeding in a clockwise fashion until all the child data items have been presented.
  • the fifth succeeding data item would overlap the first data item in the upper left quadrant and accordingly be placed in that quadrant with succeeding data items numbered 6, 7, 8, and so on overlapping successive quadrants.
  • the upper left quadrant rather than successively present multimedia data in each of the four quadrants, one may view the upper left quadrant as a starting point and an end point and present multimedia data items in a clockwise circular fashion one on top of another in a slightly stacked fashion beginning and ending with the upper left quadrant and accordingly space the multimedia data in accordance with the number to be presented in a particular group.
  • the snapshot of multimedia data items presented in Figure 12B represents the multimedia data items which may appear to a user, for example, at a particular point in time.
  • the technique with regard to presenting a particular group of multimedia data items is a layout between a parent data item and its immediate children.
  • the video title may appear in the center 1 10 as the parent node.
  • the children may be the still images or snapshots of various pictures from the movie appearing as still images 100, 102, 104, and 106 in Figure 13.
  • the title of the group is placed in the center of the canvas or presentation area.
  • the technique then rotates placement of succeeding images such as 100, 102, 104, and 106, in a clockwise fashion around four quadrants 92, 94, 96, and 98 beginning with the upper left quadrant 92.
  • the title multimedia data or parent 1 10 always remains on top of the still images and/or other multimedia data or children while succeeding multimedia data items may overlap each other.
  • Each of the quadrants itself is divided further into quadrants, and a corner of each multimedia data item is randomly placed within the outermost quadrant. That is, each of the multimedia data items shown in Figure 13 has a corner placed somewhere in one of the rectangles 92, 94, 96 or 98.
  • Such a technique for presenting multimedia data items provides a predictable temporal sequence by providing a regular rotation that is further reinforced through a layering effect in that more recent data items are always in a layer above the less recent ones.
  • the techniques and the user interfaces are widely applicable to any variety of data being presented associated with any particular domain or application in which the content may be structured hierarchically. Individual objects or multimedia data in the content may be rendered in a form of image based summary information.
  • the techniques described are capable of using a number of different spatial layout techniques and transitions for rendering individual multimedia data items to be presented in a particular presentation area.
  • the actual user controls as included in the control areas of the previously described user interfaces are also important.
  • each of the multimedia data items may be presented for a longer time period than using the prior art slide show technique improving recall and depth of processing of a particular set of objects when presented to a user.
  • the combined temporal and spatial presentation of the foregoing technique provides coherence in associating content from the same grouping of objects with each other.
  • the controls allow the user to make adjustments to the speed as well as the direction of the temporal presentation.
  • the user controls 130 may be used to control both direction and speed in which the position of the mouse cursor, for example, within the arrow is an indicator of the speed. As the mouse arrow approaches the arrow tip, the speed increases up to the maximum that the display device may support. Visual feedback is indicated to the user by the level of color saturation within the arrow. In other words, the faster the images/and or other multimedia data items are presented, the closer to the arrow tip the mouse cursor is, and the darker or deeper the color of the display arrow.
  • instant on/off is achieved when the mouse is over the user control 130 or on the user control 130.
  • the user does not need to select by depressing or clicking the mouse button in order to turn on or change the speed, or its direction.
  • Simply by placing the mouse cursor over a particular portion of the arrow causes the speed to be altered or the direction to be changed.
  • Cursor off of the arrow causes the speed to return to the predetermined rate, or stopping the presentation.
  • Elements 136 and 138 are variants of the controls of the arrow depicted in 130.
  • the slidebar 134 may be used also to slide to various points within the presentation of the multimedia data items in a continuous sequence.
  • element 136 may have any number of additional arrow icon blocks in the forward and reverse direction offering a finer granularity of control.
  • the arrangement in 136 has one block of arrows for the forward and reverse direction.
  • Another embodiment may include two or more blocks of arrows for each of the forward and reverse directions.
  • the different varieties of control buttons displayed as included in Figure 14 include a feature such that the speed and direction may be controlled by instant on off with a mouse cursor, for example, entering or leaving a display icon corresponding to the user control area.
  • the user controls determine the speed and direction of a presentation in which the movement of the mouse cursor, for example, over a spatial area of the screen starts the presentation of a sequence of multimedia data items. Movement within a spatial area corresponding to the directional arrow determines the speed of the presentation. Subsequently, movement away from the spatial area corresponding to the user control arrow stops the presentation.
  • the mouse over the arrow button area serves as a trigger for starting a temporal sequence and the mouse leaving this predetermined area causes the presentation of the multimedia data items to cease.
  • the use of two or more synchronized displays provides a traversal in browsing of multimedia data items in a hierarchical structure.
  • the presentation of the multimedia data items is generally controlled temporally with forward and backward and faster and slower controls for the purpose of content selection or assessment.
  • the multimedia-based content of the individual items in this sequence may be rendered in a designated spatial container.
  • multimedia data items are presented in a synchronized fashion with two or more components of the user interface, such as the menu selection or outline and the presentation area upon which multimedia data items are presented. This provides a technique for viewing or selecting a set of information object out of some larger collection.
  • One technique was presented for grouping objects with various structure relations, such as the parent/child relation and the 3D flyby presentation.
  • the foregoing techniques are flexible in that they may be applied to a variety of domains of multimedia data and applications. Using combinations of the controls and the selection techniques, images and other multimedia data items may be presented for efficient selection based on content.
  • the foregoing techniques may be used in applications for browsing and selecting from multimedia data.
  • One such system that includes the foregoing browsing techniques will now be described.
  • Content selection may be performed in connection with multimedia data from a variety of applications such as those using the Internet.
  • the foregoing browsing techniques may be used in indexing into multimedia data using spatial and temporal aspects.
  • a synchronized spatial- temporal technique for browsing images for the assessment of content may be used to browse and index into different varieties of multimedia data associated with the titles. Due to the temporal nature of some forms of multimedia data, such as those involving audio and/or video data, it may be difficult to "drill down" or narrow a search to a subset of large volumes of data.
  • the foregoing browsing techniques may be employed in a system that provides assistance in browsing and selecting from a subset of multimedia data.
  • the system 300 includes multimedia information 302 for which indices are created by an index extraction process 310 with an optional filtering process 312. Database records corresponding to, and representing, the multimedia indices are created 314 and stored in the index database 308.
  • the browsing tool 316 uses information regarding the indices, as read from the index database 308, to allow the user to browse various multimedia indices. Using this browsing tool 316, a user selects a particular index and control is transferred to an application, such as 304 or 306, for presenting the corresponding multimedia information 302.
  • the spatial/temporal image browser 316 shown in Figure 15 in the system 300 may be an embodiment of the previously described browsing tool as described in conjunction with the computer system 10 of Figure 1.
  • the browsing tool 316 displays various multimedia information in one of several display formats, such as those previously described, to aid a user in the selection and viewing of various portions of multimedia presentations as may be stored in the multimedia information 302. The user then selects a particular index corresponding to a particular multimedia presentation as may be included in the multimedia information 302. Once this selection is made, control may be passed to an application, 304 or 306, for displaying a multimedia presentation.
  • the multimedia data which may be extracted and used as an index for the various multimedia presentations includes any one of a variety of multimedia types, for example, such as image data, audio data, other video data, view graphs and other media streams.
  • the multimedia information 302 may include different types of multimedia presentations and information and may further include one or more varying streams of information combined together to make a single multimedia presentation.
  • one multimedia presentation may include four media streams in connection with a presentation such as a speaker that performs the presentation with view graphs.
  • the four different media may include, for example, view graphs, an audio stream, a video stream and speaker notes.
  • a second type of multimedia information that may be included in the system 300 includes video key frames that index into a video stream and an associated audio stream. This is an example of a video clip having a corresponding audio track, for example, as when viewing movie trailers or clips.
  • Such multimedia information and resentation may be used, for example, in connection with a video-on-demand application.
  • a third type of multimedia information that may be included in the system 300 may be related to consumer online shopping, as may be included in an electronic commerce application providing product descriptions and information to consumers shopping over the Internet.
  • This type of multimedia presentation may include several different types of media streams similar to the video-on-demand. For example, there may be a presentation or a video clip for a product demonstration as well as additional text and audio information, such as a jingle or a verbal description, placed in an audio file accompanying the product demonstration. Each of these items may be included in a different media stream which, when in combination, result in multimedia presentation for a particular product.
  • flowchart 330 are method steps relating to the data flow and description previously set forth regarding the system 300.
  • one of the media types associated with the presentation is chosen as the indexing media.
  • view graphs may be used an index.
  • a time stamp may be used to synchronize a particular point in the video, audio and speaker notes accompanying the corresponding view graph.
  • Each view graph may be associated with a time stamp and a corresponding point found within each of the other media streams.
  • a video key frame may be used an index into a corresponding audio stream.
  • a product image which represents a key frame in a video such as a product demo may be used as an index similar to the video key frame previously described with regard to a movie.
  • the video key frame representing a product image is similarly used as an index into accompanying multimedia streams associated with a particular product being represented in a multimedia presentation.
  • image-based indices are generated in accordance with the indexing media. It should be noted that at step 334, if other types of indices besides those which are image -based are used, these are also generated. It should be noted that at step 334 the image-based indices which are chosen are those which will be displayed in a particular format, for example, as by the browser tool 316 in one of the previously described display layouts to allow a user to select a particular multimedia presentation in accordance with the displayed multimedia indices. It should also be noted that different processes for generating the image-based indices at step 334 may vary in accordance with the particular media streams as well as the variety of software that may be available in each particular implementation.
  • Example embodiments and processes for generating different types of indices will be described in more detail in paragraphs that follow. However, it should be noted that this should not be construed as a limitation as to techniques used to generate indices which are image-based used in conjunction with multimedia information 302.
  • selection of particular portions of a medium may be done in an automated fashion as well as a manual fashion, or in combination.
  • a video frame may be a video key frame in accordance with the amount of change detected between two consecutive images in a video. This may be done in an automated fashion, for example, using software which determines this difference in accordance with the data type and structure of the video files.
  • Another key frame selection technique that may be performed, such as using videos with regard to the multimedia shopping presentation, may be performed manually.
  • a video key frame may be selected as one to be used as an index, for example, if that frame of the advertisement for a product includes all of the components of a product in clear view. This may be performed manually by a person viewing the advertisement or clip for a particular product. Regarding the multimedia presentation of the speaker, each view graph included may be used as an index medium.
  • Step 336 optional image filtering processing is performed to remove any extraneous images.
  • Step 336 may be performed in an automated or a manual fashion.
  • a person may view all of the image indices associated with a particular multimedia presentation and determine that one should be removed for any one of a variety of reasons in accordance with predetermined criterion. For example, a threshold number of indices may be chosen and a decision may be made to remove the first and the last image index for a particular presentation.
  • an automated process may be performed upon the selected image indices associated with a particular multimedia presentation to remove one or more images that do not meet predetermined criteria.
  • Such predetermined criteria for example, may include images of a particular color, including certain objects, or other criteria.
  • One or more indices may be removed or "filtered out” in connection with, for example, heuristics related to the amount and/or number of colors, such as concentration and contrast related to image coloring, degree of similarilty or difference between two images, and temporal proximity.
  • a structured index object is created that may be stored in the database 308.
  • one or more records are generated in accordance with the various indices.
  • the multimedia presentation as may be included in multimedia information 302 is updated if needed. For example, at step 340, a time stamp associated with each view graph of a multimedia speaker presentation may be determined. It may then be necessary to accordingly update one or more of the associated multimedia streams also associated with that presentation to synchronize the multiple media streams of the presentation. This enables, for example, a synchronization point in each of the media streams for a particular view graph. Similarly, this update step may be performed for other types of multimedia presentation, such as the video and audio streams associated with the video-on-demand selection. Step 340 may be performed prior to this point in processing as included in the flowchart 330.
  • step 336 it may be beneficial to perform this step subsequent to performing any optional filtering so as not to unnecessarily update other media streams in the multimedia information 320 for indices which may be removed by the optional filtering step 336.
  • Each technique for updating the different streams in accordance with, for example, the indices selected may vary with each particular embodiment and implementation. Particular examples will be described in paragraphs that follow.
  • the index database 308 may be updated to include the information for the indices created or determined for each particular multimedia presentation.
  • flowchart 330 is the process by which indices are extracted from the multimedia information 302 and information may be stored in the database 308 corresponding to the various multimedia indices.
  • the multimedia information 302 for which multimedia indices are determined may include, for example, an entertainment movie, a seminar style presentation, and an online shopping presentation, as well as other types of multimedia presentations in accordance with applications.
  • the indices may be presented for viewing in accordance with a predetermined structure or format.
  • an HTML file may be dynamically generating using information from the index database for each multimedia presentation to be presented in accordance with a user selection, such as seminars or speaker presentations may be displayed in accordance with a predefined viewing format for a particular speaker selected by a user.
  • the format or viewing structure of the various titles may be predetermined.
  • the HTML file may be dynamically generated in accordance with the particular user selection.
  • the record 350 is an example of a database schema record that may be stored, for example, in the index database 308 of the system 300 of Figure 15.
  • the record 350 in this embodiment includes three fields. It should be noted that although what will be described are three fields, other embodiments may include records that include a varying number of fields with other additional information in accordance with each particular implementation.
  • the record 350 includes a pointer to the image to be displayed by the browsing tool 316 in the field 350a. Also included in the record field 350b is an index or time stamp reference into the multimedia presentation.
  • Filed 350c specifies a software application to which control is to be transferred.
  • field 350c includes a URL with a file extension.
  • the URL may identify a multimedia presentation and the file extension may identify a software application that is invoked to present the multimedia presentation. It should be noted that other techniques may be used to both specify the software application and the multimedia presentation in accordance with the functionality available with each particular embodiment.
  • Field 350a which is a pointer to the image to be displayed by the browsing tool 316, may be a file identifier or handle identifying an associated image, such as a file identifier of a JPEG or .GIF image file.
  • the information 350a is associated with or represents an address of an image that is displayed, for example, such as an image included in the display format 182 of Figure 11 such as images 160 or 162.
  • the index or time stamp represents a reference with regard to the image associated with 350a.
  • the field 350a may be a pointer to an image which is a portion of a video in which the video key frame associated with field 350a occurs at a particular point in time with respect to a video clip. The point in time is referenced by field 350b.
  • This index may be, for example, a temporal offset into a video with regard to playing time from the beginning of the clip.
  • Field 350c represents a software application which, if the associated image is selected at field 350a, is invoked to display the multimedia presentation. In other words, field 350c is associated with a software application to which control is transferred to present one of a variety of different multimedia presentations.
  • the record format 350 of Figure 17 includes information which is displayed by the browsing tool, such as the images associated with fields 350a.
  • a user selects one of the associated images displayed, and accordingly, a software application is invoked through a control-transfer mechanism in accordance with each particular embodiment.
  • the particular software application is identified in field 350c.
  • the software application may be transferred control along with additional information such as the index time stamp reference into the multimedia presentation represented by field 350b.
  • FIG. 18 shown is a more detailed description of a particular database schema record previously described in conjunction with Figure 17.
  • information may be included in the record in accordance with the previously described record structure for a particular multimedia presentation.
  • Record includes an identifier of a file of a view graph miniature image in the first field, with the time stamp in the second field, and a software application identifier of a multimedia presentation player in the third field.
  • the record 362 in the view graph miniature image file identifies a file, such as a URL or a file name, in accordance with a particular file system of the image that is to be displayed by the browsing tool 316.
  • the second field includes a time stamp which represents the particular point in the presentation and in the associated multimedia streams which correspond to the particular image represented in the first field which has been chosen as an index into the presentation.
  • the third field of record 362 is a presentation player URL.
  • a presentation player which has the name of the DejaView software application is identified by a particular file extension on a URL that may be included in this field. For example, a URL with the file extension ".DJVW" specifies that control is to be passed to a particular software application.
  • the third field may include a URL describing a particular presentation in which an association may be made between a software application and a particular file extension.
  • the third field may specify a URL of a file with a file extension associated with a particular software application through a MIME type.
  • the browser may use an operating system supplied control mechanism for passing control from the browser to the specific software application.
  • an operating system supplied control mechanism for passing control from the browser to the specific software application.
  • the particular mechanism by which control may be passed is in accordance with each particular embodiment and implementation and operating system-supplied mechanism.
  • the first record 362 may be an index into a speaker presentation as previously described which includes view graphs as well as speaker notes and a video and audio stream corresponding to a speaker presenting the view graph.
  • Record 364 may correspond to an index, for example, with a video key frame serving as an index into a movie.
  • the first field of record 364 identifies an image which is a video key frame.
  • the second field represents a time stamp which is a temporal spacing into the associated video at the position where the video key frame occurs.
  • the third field identifies a pay-per-view application, such as by a ".PPV" file extension which result in invoking a pay-per-view application.
  • images displayed with this associated index may be a series of indices into movies or videos displayed, for example, by the browser tool 316 when a user is selecting a movie for a pay-per-view application.
  • Record 366 may correspond to an index, for example, of a particular product of an e-commerce shopping application.
  • the first field may be associated with a file that is a product image.
  • the second field may represent a time stamp or index into a video, for example, corresponding to a product demonstration.
  • the third field may include a URL to an e-commerce shopping application identified through a particular file extension and its corresponding MIME type.
  • the foregoing is just one example of an embodiment of a database schema and record format which corresponds to each of the various indices in accordance with each of the multimedia data types. These records may be created and stored in the index database 308.
  • the database itself may be any one of a variety of types of database such as an object database or relational database that can be implemented with any one of a variety of commercially available packages.
  • information about the indices have been gathered and created and stored in database 308. What will now be described in the paragraphs that follow is how this information stored in the index database may be loaded and used by the browser 316.
  • step 372 a subset or reduced set of data has been obtained in accordance with user search criteria.
  • step 372 has the purpose of reducing the candidate set of data for which information is to be displayed by the browser 316 to a reasonable subset, for example, that may be displayed in accordance with one of the previously described display layouts and techniques.
  • This subset of information may be obtained using any one of a variety of existing search engines and techniques. For example, for each multimedia presentation included in the multimedia information 302, categories may be associated with each of these.
  • Each of the multimedia presentations and corresponding information may be indexed and stored in some type of structure from which data may be stored and retrieved, such as a linked list or other data structure known to those skilled in the art.
  • a user may perform an initial query in accordance with predetermined categories associated with each of the multimedia presentations.
  • the names of various multimedia presentations may be the resultant set.
  • the results set may include handles or identifiers associated with each of the records such as 362, 364, and 366 stored in the database 308.
  • the embodiment of Figures 17 and 18 in the database schema record include, for simplicity, only that information which is associated with each of the applications and images.
  • This information may be stored internally in the database 308 within each of the records, or separately in a different data structure or database schema. For example, if the category is a comedy movie, stored in the database 308 or elsewhere may be a list of database records associated given titles with predetermined categories. This may, for example, save time when performing a query of the database by having these records already pre-selected and predetermined for each of a particular category.
  • a database query is performed to retrieve the records corresponding to indices for the subset of data to be displayed by the browsing tool 316.
  • the information retrieved as a result of step 374 may be the image URLs and other associated fields stored in each of the records.
  • a hierarchical data file is dynamically generated which represents these query results.
  • the URLs may be organized in a format, for example, such as the hierarchical data file 150 of Figure 2A in accordance with a predetermined display format. In other words, the URLs may be arranged in a particular format or structure in accordance with the display or viewing format.
  • the image indices are displayed using the browsing tool 316 in connection with the hierarchical data file.
  • one of the displayed image indices is selected by a user such as from a display area having one of the image display formats, for example, as shown in Figure 9 or Figure 1 1.
  • control is transferred at step 382 to a software program or application which presents the multimedia presentation corresponding to the index selected.
  • the browser 316 may be used to display the indices. A user makes a selection.
  • the image browser 316 uses the information from the third field in the record, such as record 364, to invoke an application such as the pay-per-view application.
  • a particular multimedia presentation for example, as stored in accordance with the subset of information at step 374.
  • This data structure may be stored, for example, in the form of lists predetermined as previously described.
  • the multimedia information is obtained from the multimedia information block 302. It should be noted that the browser 316 may obtain additional information as needed from the multimedia information store 302 in order to transfer control to a software application.
  • the software application to be invoked is represented in the third field, for example, of a database schema record
  • control may be transferred to another application with the appropriate parameters and other information to invoke and display the multimedia presentation corresponding to the selected image index.
  • the techniques available for performing a database query for example, of the database 308, may be supplied with a commercially available database package that may be chosen in accordance with each particular embodiment.
  • Application Programmer Interfaces APIs
  • Information as to which of the indices correspond to a particular multimedia presentation associated with the subset as used in step 374 may be stored in the database 308 or elsewhere.
  • the data structure 420 represents a relation between a predetermined set of categories and associated multimedia presentation items that are movies. Similar data structures and others may be used to represent this and other types of multimedia information in accordance with each of the applications, such as 304 and 306.
  • the data structure 420 includes a list of predetermined categories. Each category may be associated with one or more multimedia items. In this example, cat-1 422 is associated with movie 424 and movie 426.
  • the categories may be a set of predetermined categories in which, using either automated and/or manual techniques, categories are mapped to multimedia items, such as movies 424 and 426.
  • a user may be selecting a movie in connection with a pay-per-view application. The user may select a category of movies, such as comedy, that corresponds to one or more categories, such as cat-1
  • the user query may result in only a single category, cat-1 422, and only those movies associated with this category. It is from this list of movies that the user may select a single movie or video resulting in control being passed to a video-on-demand or pay-per-view application program.
  • Data structure 430 may be used in conjunction with data structure 420 to display one or more indices using the browsing tool 316 in connection with a user query. In this example, once one or more movies may have been selected using data structure 420. Data structure 430 may be used to obtain the one or more indices associated with each movie. It should be noted that other data structures besides 430 may be used to store and retrieve information as needed to display associated indices using the browsing tool 316. For example, a linked list rather than a table format may be used. Other variations known to those skilled in the art may be used with appropriate storage and retrieval techniques in accordance with each variation.
  • the information represented in data structure 420 and 430 in combination relates to one embodiment in which the multimedia presentations of interest to the user are reduced in accordance with predetermined categories associated with each of the presentations.
  • Other embodiments may use other techniques to reduce the number of multimedia presentations of interest to a user.
  • the data structures used in connection with these techniques may also vary. The storage location of these data structures may also vary in accordance with each embodiment and system configuration.
  • FIG. 22 shown is an example of an embodiment of a user interface display 400 to display video titles using the browser tool 316.
  • This may be an interface, for example, presented with a video-on-demand selection as a service presented with a choice of entertainment video title where a user selects a particular video that they desire, for example, to have streamed over IP networks or other networks to home PCs or other display devices. Due to the availability of large numbers of titles available from the network, users of the service require support for narrowing down the space of potential selections.
  • a particular subset of category selection may be performed resulting, for example, in a display of Figure 20 , 400, in which video key frames from designated movie favorites are presented as dynamic collage.
  • the user may select, as by a mouse click, a key frame of interest which results in the transfer of control to another application such as 306 to a new screen.
  • FIG 23 shown is an example of an embodiment of the screen displayed when the selection of a movie title has been made from the screen 400 of Figure 20 using the browsing tool 316.
  • the display 402 of Figure 23 presents a movie trailer and other information about an associated title selected.
  • the user is assisted in "drilling down quickly" through a large space of movie titles to a point where a choice of whether to select or pay for a movie delivery can be reached.
  • techniques by which representative video key frames may be generated from video content either automatically or semi- automatically are well-known to those schooled in the art.
  • server software may dynamically generate an HTML file that represents the results via a query resulting in the title skimming screen, for example, as shown in conjunction with Figure 22 400.
  • the browsing tool 316 takes as input the HTML or XML file that specifies the sequence of title objects, each of which is associated with a sequence of images that may be specified, for example, as URLs previously described in conjunction with the hierarchical data file 150 of
  • FIG. 2A Also as previously described, as part of the off-line processing, previously stored images, such as JPEG and GIF format files have been made available and accessible using various protocols such as HTTP. As the user views the title sequence, images are loaded incrementally in conjunction with the various display formats chosen. Any one of a variety of different layout techniques may be used, such as those previously described in conjunction with other figures herein.
  • this DejaView player is a software application that runs as a client on a user's computer in which the user experiences, for example in this application, a virtual auditorium- style presentation that includes temporally synchronized multimedia elements, such as the four media streams previously described, including view graphs, user notes, and audio and video streams of the speaker.
  • the DejaView player such as 304 included in the system 300 of Figure 15, is a tool that allows the user to navigate around a presentation in a random access fashion while maintaining synchronization between the various media streams.
  • the DejaView application may use temporal and spatial browsing techniques of the browser 316 in which view graph miniatures serve as the indices displayed by the browser 316.
  • the view graph images may be generated and stored prior to processing by the browsing tool 316.
  • control may be transferred to the DejaView player 304 to a particular presentation.
  • Figure 22 is an example of an embodiment of the user interface display of the multimedia presentation using the browsing tool in which miniature view graphs as part of a speaker's presentation are used as indices and displayed.
  • runtime or execution control may be transferred to the DejaView player 304.
  • Figure 25 shown is an example of the embodiment of a screen displayed when a selection of a multimedia presentation has been made using the browsing tool such as by selecting the miniature view graph 404a.
  • Figure 25 displays an example of a screen under the control of the DejaView player 304 when a particular view graph is selected. The user is presented with the view graph in an enlarged form. Additionally, using the time stamp information associated with the view graph, the user is fast-forwarded to different places corresponding to the time stamp which are synchronized with the time stamp in the associated media streams. In other words, view graph 404a which has been selected occurs at a particular point in the presentation associated with a first time stamp.
  • the first time stamp represents a marker or a "book mark" into the other video streams: the video stream, audio stream and the like.
  • the DejaView player provides, for presenting to the user, a point of synchronization in each of the media streams.
  • Figure 26 shown is an example of an embodiment of a user interface display of a multimedia shopping title in which the titles are displayed using a browsing tool 316.
  • the proposed technique for indexing into multimedia presentations includes online all e-commerce shopping. Vendors such as QVC have demonstrated consumer appeal for convenience shopping services through the likes of QVC's product showcase television products.
  • FIG. 27 shown is an example of an embodiment of a screen display when the selection of a multimedia shopping title 408a has been made using the browsing tool 316
  • the screen displayed m Figure 27 is that associated with a particular other application such as 306 where control has been transferred from the browser 316
  • a DejaView Player 304 is one of the applications that may be used to display, for example, a multimedia presentation selected by a user in conjunction with the browsing tool 316 What will now be described is one
  • the DejaView player 304 is part of a software application that may be used to capture and play multimedia information In particular, this may be used for real time capture and creating synchronized multimedia streams therefrom which may accessed using a communications connection, such as one of a variety of network connections
  • the capture of the multiple multimedia streams may generally refer to media journaling techniques for capturing events, such as a lecture style presentation with visual aids m real time as they occur and then processing the combination of streams to create a synchronized multimedia presentation.
  • the DejaView player 304 may be described as part of a media journaling application.
  • FIG. 28 shown is an example of an embodiment of a media journaling application that includes the player 304 and other application components.
  • This journaling application in functional terms, provides a user with the capability of capturing multimedia content and related information to create a synchronized multimedia presentation. For example, a seminar style presentation may be captured digitally in real-time. Using certain post-processing techniques, within a very short period thereafter, additional information is combined and synchronized therewith. The new presentation is then made available on a network such as the Internet through a media journal player.
  • the player 304 is, for example, a web browser-based and allows the user to view the whole presentation, or allows for seeking to different places in the presentation while maintaining the synchronization.
  • a media journaling system 1100 which includes an audio/video capture digitizer 11 10 that works within a master capture 1120 to create and capture an audio/video presentation for further combination with other inputs, for example, a viewer's notes or a presenter's slides.
  • the audio/video master capture 1110 includes an audio/video capture 1112 which is adapted to receive, for example, an analog video or audio signal, or both (hereinafter referred to as an audio/video signal).
  • the captured audio/video signal is converted to corresponding digital signals that are further compressed in the audio compressor 11 14 and video compressor 11 16.
  • the audio/video capture and digitizer 1110 are well known in the art, and are commercially available, for example, as a Winnov
  • the master capture 1120 includes a camera/microphone 1122 for actually recording an audio/video event.
  • the output of the camera/microphone 1 122 is typically an analog output that would require digitization and compression by the audio/video capture and digitizer 1110.
  • audio/video interleaved (AVI) output may be immediately available such that the audio/video capture and digitizer may be bypassed in its entirety.
  • Time stamps may be determined, for example, in an automated or manual fashion for each of the viewgraphs in which each time stamp is a temporal marker in the speaker presentation.
  • this time stamp recordal may be automated using a user interface as may be included, for example, in the journaling tool 1 124. Using this tool, the user may select a particular viewgraph to be used as an index. This selection may be made, for example, by clicking a mouse button. Upon this selection, corresponding time stamp information may be recorded in a predetermined file format.
  • One or more of these timestamps may be stored in a timestamp file.
  • the timestamp file may be used in connection with an AVI file, for example, which interleaves audio and video information into a single file format.
  • An AVI file may be produced, for example, by monitoring real-time capture of raw audio-video feed.
  • an NTSC analog video feed from a camera may be connected to a personal computer.
  • the personal computer may be, for example, a Pentium-class processor having a video digitizer card to digitize the incoming audio/video stream in real time to produce a single interleaved stream, an AVI file.
  • other information such as textual notes, may also be generated using similar manual and/or automated techniques.
  • the journaling tool 1 124 used to record the presentation information may be, for example, incorporated into a standard web browser such as Microsoft Internet Explorer or Netscape Navigator. As known to those skilled in the art, alternatives, such as a stand-alone solution, may also be used in which standalone applications may access the Internet or other network directly.
  • the journaling tool may use, for example, Windows Media Technologies from Microsoft Corporation, to integrate timestamp information that may be recorded also using the journaling tool as previously described.
  • Each of the different media streams may be indexed with the time-stamp information, as may be gathered and stored in the time-stamp file.
  • a commercially available Microsoft tool vidtoasf
  • This tool outputs such a file in a commercially specified file format, the ASF or Active Stream File format that may be served or provided by a software server, for example, written using Microsoft Windows Media Technologies.
  • the ASF file has the time stamp information embedded in it such that a DejaView player may trap and respond to viewgraph change events as the ASF stream the player is receiving plays out.
  • the player interprets these events using scripting logic within DejaView that updates the slideshow, notes and controls of the tool.
  • the ASF file may be received by the DejaView player.
  • the Windows Media Player ActiveX controls used by the player to present the stream is able to raise events corresponding to the timestamps detected in the stream. These events may be trapped and interpreted by scripting logic included in the player.
  • ActiveX controls are generally reusable pieces of code written for example, in C++, Java, Visual Basic and the like. These may be incorporated into an HTML page to extend browser capabilities.
  • Microsoft makes available a variety of ActiveX controls, such as those installed with Internet Explorer. Controls used in one embodiment of DejaView, for example, include Microsoft's Windows Media Technologies Control and Image Control.
  • the Windows Media Player ActiveX control functions as a player of streaming audio/video, embedded in a Web page.
  • viewgraph images may be displayed using the Image ActiveX control using, for example, the PicturePath property which is the URL of a .GIF or .MPEG image to be displayed.
  • journaling tool may either be physically co-located or physically remote to the journaled event location.
  • the journaling tool may include a digitizing PC, video source (from camera), and audio input (local microphone).
  • Available video capture software located on the digitizing PC includes VidCap from Microsoft Corporation (which does not compress audio) or Premiere from Adobe.
  • the journal event's audio/video feed may be supplied in analog form via a cable, or digitally via a data network, for example.
  • the ASF file may be produced including the integrated indexing information
  • the ASF file is in a form that may be streamed from a Windows Media Technologies server in real time to the journaler player, the DejaView player 304
  • the ASF movie may be stored in the Windows Media Technologies server 1 140 along with other media on a web server and is ready for streaming from the server 1 140
  • the Windows Media Technologies server is currently available from Microsoft Corporation
  • the ASF movie includes the timing information, as may be recorded using a journaling tool
  • a journaling playback device, such as the player 304, may be able to trap and respond to viewgraph change events as they occur accordingly updating other data streams, such as notes
  • the slide miniatures may be located, for example, also on the server 140 that may optionally be connected to a journaling tool
  • Viewgraph miniatures may also be included m this embodiment, for example, to be displayed by the browser tool 316
  • These miniatures may be created using Microsoft Powerpo ⁇ nt97 if the viewgraphs are created using this application also This may be done, for example, by using the the Powerpo ⁇ nt97 "Save as HTML" wizard to generate the JPEG images
  • the journaling software application may include a player and a server component
  • the DejaView player 304 may be the client portion of the application residing and executing on a user's computer
  • the DejaView Player may be written, for example, using Windows Media Technologies and other components as described elsewhere herein
  • the DejaView Server portion may be written using the commercially available Microsoft Windows Media Technologies Server, for example, in which the server computer may be running Windows NT Server 4 0
  • a Pentium class PC running Windows 95 or Windows NT 4 0, for example may be used
  • a video digitizer card may be, for example, the Winnov Videum AV card that also supports audio capture
  • a video camera, such as a SONY H ⁇ -8 CamCorder may be used to capture the audio/video feed of a presenter or speaker
  • other cameras capable of producing an NTSC video out may also be used. It should be noted that the NTSC out of the camera may be fed directly into the video input of the video capture card of the computer
  • the DejaView Player 304 may run, for example, on a Microsoft Windows 95 or NT platform and operates in one embodiment using Microsoft Internet Explorer
  • Playback of a DejaView presentation may be used in a computer system in conjunction with the Microsoft Windows Media Technologies Player.
  • the Windows Media Technologies Player ActiveX control is an integral component in one implementation of the DejaView player.
  • the Microsoft Windows Media Technologies Server may be used to stream audio/video content to DejaView players.
  • the Microsoft Windows Technologies Server is supported, for example, on an NT 4.0 server platform.
  • Video capture software tools may be used to digitize DejaView presentation feeds.
  • Microsoft VidCap or Adobe Premiere may be used.
  • different hardware such as the Videum video capture hardware, may also include software utilities for this purpose.
  • the ASF file may be made available to clients from a Windows Media Technologies server.
  • Other relates files, such as the JPEG viewgraphs, may be placed in locations on a Webserver being used in support of the DejaView player
  • the player 400 is an example of an embodiment of the DejaView player 304. It is depicted implemented as a web-based application, and hence is shown as accessed from a predetermined web site 1401.
  • a web command window 1402 is the web based application interface, i.e., a Netscape Navigator or Microsoft Internet Explorer interface for allowing a user to issue commands via the web interface.
  • the journal player 304 allows a user to playback and interact with multimedia presentations of joumaled events. As the ASF movie is received by the journal player 304, Windows
  • a talking head window 1403 provides a video window that presents a talking head video of a presenter (with accompanying audio), for example
  • a graphics display window 1404 displays graphic images, for example, the presenter's view graphs
  • An electronic notepad window 1421 displays, for example, textual information pertaining to a presentation
  • a user may also create additional notes while viewing the presentation when a timing event coordinator tool is provided (to be described) Such newly created notes may be kept m synchronization by the journaling tool
  • the notes are storable on network servers for subsequent retrieval, or for sharing amongst users (several sets of notes) in a groupware environment
  • the journal player 304 may present the contents of the talking head window 1403, the graphics display window 1404 and the electronic notepad 1421
  • the journal player 304 may also include a plurality of user controls for allowing a user to navigate the player to different parts of a presentation
  • the first set of user controls are grouped together as navigation controls 1410
  • Navigation controls 1410 include a play control 1412, a pause control 1413, and previous slide 1414 and next slide 1415
  • the play control 1412 causes the presentation to begin to play, either from an initial start of resuming from a pause position
  • the pause control causes the presentation to temporarily freeze until the pause control 1413 is again selected or the play control 412 is elected
  • the play and pause controls 1412 and 1413 are mtegratable into a single control that toggles from the play to pause function and vice versa according to the current function
  • a table of contents menu 141 1 is also included that provides a menu displaying the titles of the view graphs used by the presenter, which menu effectively acts as a table of contents for the presentation A user may thus browse a presentation and go directly to an area of particular interest by selecting the view graph
  • a second set of user controls are the browse controls 1420 that allow a user to browse the user notes that are displayed in notes window 1421. There may be a one-to-one mapping of the user notes and view graphs in the journaled presentation as ensured by the journaling tool.
  • the note window controls 1422 through 1426 the user may move back and forward any number of note pages independently of the rest of the presentation. Forward note control 1422 and back note control 1423 allow the user to go forward and backward through the notes, respectively.
  • journal player may cause the journal player to resynchronize the rest of the presentation with that note by selecting a seek control 1424.
  • a new note control 1425 is provided which allows the user to enter a new note, that may be synched to the presentation at its current position upon selecting the save control 426.
  • An AVI movie may be created in step 510. This step may additionally be an audio only capture, a video only capture, or an audio/video capture. Alternatively, the AVI movie may take the form of selecting and arranging content from pre-existing sources, such as a video tape of an event.
  • time stamps may be captured for any information related to the AVI movie of step 510. As already described, view graphs may be time stamped according to their presentation with a speaker (AVI movie).
  • a video stream is generated at step 530 with the time stamps now encoded in the AVI movie
  • the images or view graphs are elected in step 540 so that during a playback these images may be located and co-displayed with the presentation
  • these selected viewgraphs may be the multimedia indices, for example, stored in the database 308 and displayed by the browsing tool 316
  • the images and any other items for playback of a presentation are stored on appropriate servers
  • the playback of the presentation on a journal player is accomplished at step 560
  • the player 304 may be transferred control by the browsing tool 316 upon the user may requesting a presentation, for example, from a list
  • the images, notes and other information of the presentation may be included in the multimedia information storage 302 This information used in connection with the presentationsmay be accessible from a web site and the information accordingly provided to the journal player with the images, notes, and other non-presentation information being downloaded to the player from their
  • Timing event coordinator keeps time and coordinates events among software componentsthat exhibit temporal state changes.
  • Use of the timing event coordinator herein will be described in an environment for synchronizing components within a multimedia presentation.
  • a presentation typically engages a user, has some initial state, T eg j n , and changes state at discrete times as the presentation progresses to an end time, T end -
  • the coordinated components are the separate pieces of the presentation (events) that used to change state at discrete times.
  • the timeline 601 is an example used in a presentation having only two coordinated components, A and B. However, the following may be generalized to more than two components and associated events.
  • the timeline 601 includes numbered state changes for each component.
  • component A may represent view graph change events having an initial state Ao at the start of the presentation and first changes state at A] .
  • Component B may represent note change events, for example, and may have an initial state B 0 and first changes state at Bj .
  • a synchronized application may be created by having each component maintain its own time relative to the start of the presentation, T R EL- Alternately, each component may rely on a dedicated timing coordinator for notification to change states. This approach is described in more detail below.
  • the timing event coordinator may maintain state information for each component for the set of times, S, each relative to the presentation start time, T REL , as to when they should change their state.
  • S ⁇ A 0 , Ai, A 2 , A 3 , A 4 ⁇ .
  • Each type has an index associated with each time, which index identifies the sequential relationship or order (the first state for that type, the second state, etc.). Additionally, each type may have a flag associated therewith to provide information as how to treat that type.
  • the flag may be binary state of a "0” or a "1" that differentiates between two different treatments in which "0” means treat special, and "1" means no special treatment.
  • Special treatment in this example, may be that if a user jumps to a certain location in a presentation, then that component type with the "special flag" set may be treated so as the most current component relative to that location, should be displayed.
  • the timing event coordinator may merge S into a table of triples having value, type and index for each element of
  • the first two triples of C describe components A and B respectively and refer to their initial values at time zero, with A represented as a type one, and B represented as a type two, each having an index initially at zero.
  • the third triple in C is associated with component A for a time equal to five seconds, and has an index of "1" (the second A in the set).
  • the timing event coordinator supports an interface that allows a component to set the time T RE L to an arbitrary time value, V. If this occurs, the timing event coordinator looks in C for a matching time value. If a match is found, events are generated with the corresponding type and index fields. These events are used to instruct the component represented by the associated type to perform that action at that time. If no matches are found, the timing event coordinator looks to the next greater triple in C. calculates the difference between the current T REL and sets a timer to raise an event at this time difference. Each time and event is raised by the timer, the period to the next event is determined for the next greater time value and the timer is set accordingly. This is repeated for each next greater time value.
  • a reference for the timer concerning where in the video stream the event occurs is maintained by creating a series of predetermined time stamps in the video stream.
  • a time stamp may be created every second with no component or event associated therewith.
  • the granularity of such predetermined time stamps is determined according to the specific application requirements.
  • the timing event coordinator may be implemented as an
  • ActiveX control and may be used within Internet based applications including software components based on Java applets or ActiveX controls. Neither Java applets or ActiveX controls are intended to be a limitation in an embodiment or implementation of the timing event coordinator.
  • the timing event coordinator merges the notepad component registered timestamp information with the previously registered view graph display component time stamp information.
  • the timing event coordinator responds by generating two events, one to the notepad component requesting a change to note three, and one to the view graph display component requesting a change to the view graph that present at this point in the presentation. Additionally, the application invokes an operation on the timing event coordinator to get the current T REL at the note three point. This is used to inform the audio/video component where to seek to in the associated media stream. Unless otherwise directed by the user, the application proceeds in normal fashion through the remainder of the presentation. Using the timing event coordinator, users may be able to make new notes at runtime (in addition to those already recorded).
  • a user viewing a presentation may take a note at any given point in the presentation.
  • the notepad component may then contact the timing event coordinator to retrieve the current T REL at the new note point.
  • the notepad component may notify or signal the timing event coordinator and re-register its timing information using the same type as before.
  • This new note then may function as the other notes.
  • other users may make their set of notes available, for example, from a server, such that a group of users may share notes.
  • FIG 32 shown is a flowchart of method steps of an embodiment for synchronizing interrelated components of a multimedia presentation.
  • steps that may be performed by a timing event coordinator are shown in flow diagram form for synchronizing interrelated components in a presentation as described above.
  • the timing event coordinator method begins at step 701 which represents either the start of a presentation, or is caused by a user request to go to a certain time or event.
  • step 703 a test is made to determine whether any of the registered software component types have a special flag associated with them. Thus, if any one of the software components has a special flag, the result of the test will always be yes and control will flow to step 705. Assuming that a special flag is associated with one of the software components, a second test as to whether any special processing is needed is made at step 705.
  • An example of special processing includes determining whether one or more software events need to be updated (or not updated) to the requested command and its associated time. A certain application of a joumaled presentation may require that a view graph be updated to the requested time or command, but not the associated note, hence requiring special processing. Step 707 carries out such special processing.
  • step 703 If no special flags were detected in step 703, processing goes directly to step 709 where a test is made for matches in the relative time, T RE L, and the software components. A lack of any software component matches causes processing to proceed to step 71 1 wherein the presentation is restarted at T REL and a check is made to determine whether there is a greater time value, V, left in the presentation. A greater time value V will be treated in step 713 such that the difference between V and T REL is calculated and the timer is set to expire in that calculated time period. Control then returns to step 701. If there is not a greater time value, V, then the playback is concluded.
  • step 719 is executed where an event is generated such that the index for the type and index for that matching value.
  • Step 719 is performed if more than one match was found (i.e., a note and view graph changed simultaneously). Steps 717 and 719 are repeated until all matches are processed before control is given to step 711.

Abstract

Described are a variety of techniques for viewing multimedia-based information in accordance with user controls for the speed and direction of the multimedia presentation. Different techniques for presenting multimedia information (302) are disclosed. User controls providing control for speed and direction in accordance with a user selection device being in a particular location. The speed and direction controls are on when a user device is within the defined area and are off when it leaves the defined area. Also described is a system using a browser tool (316) for displaying multimedia information (302). Multimedia information (302) are indices into multimedia presentations. A user selection is made by viewing the multimedia data items and control is transferred to an appropriate application to present a multimedia presentation corresponding to a selected multimedia data item. A technique is disclosed for generating an index database (308) of the multimedia data items used as indices into the multimedia presentations.

Description

SYNCHRONIZED SPATIAL-TEMPORAL BROWSING OF IMAGES FOR SELECTION OF INDEXED TEMPORAL MULTIMEDIA TITLES
Technical Field This application generally relates to computer systems, and more particularly to computer executed techniques for locating and selecting information.
Background Art
Computer systems may be used in a variety of applications and environments. For example, in commercial applications of computer systems, a computer may assist a user in making a decision or choice based on information that is viewed using the computer. A computer storing information may be used to display information to a user. Subsequently, the user may make a decision or selection in light of the information presented. One type of information from which a user may make a selection is multimedia data.
When selecting and searching for items within multimedia data, one concern is what to utilize as an index into the multimedia data. One technique uses video key framse as an approach to indexing or accessing into portions of multimedia data. For example, video key frames may serve as reference framse and be used to index into videos. However, a problem may be encountered when the key frames are similar in that a user is associating a key frame with a distinct portion of a video segment. Additionally, when there are multiple media streams associated with one presentation, it may not be clear which one of the media streams to use as an index. In this instance, for example, use of video key frames as an index into portions of multimedia data for selecting and browsing may not be adequate.
Another problem is how to display multimedia data items and associated indices to enable one to browse and select a corresponding presentation. Browsing and selecting techniques, such as those employed by an Internet browser like Internet Explorer or Netscape Navigator, may be used in conjunction with static data or content information. However, these techniques may not be particularly suitable for use with temporal based multimedia information such as the foregoing video information and other types of mulitmedia data that may be used as indices into multimedia presentations being browsed.
Thus, there is required a technique for efficiently locating, browsing, and indexing multimedia information.
Brief Description of Drawings Figure 1 is an example of an embodiment of a system;
Figure 2 is an example of a block diagram of software components included in the system of Figure 1 ; Figure 2 A is an example of one embodiment of the hierarchical data file of Figure 2;
Figure 3 is an example of an embodiment of a block diagram illustrating the flow of control in producing and using various software components included in the system of Figure 1 ; Figure 4 is a flowchart of an example embodiment of method steps performed at run time when a client browser 16 issues a request for the software browsing tool; Figure 5 is an example of an embodiment of the browsing software tool that may be included in the system of Figure 1 ; Figure 6 is an example of a user interface display for a generated browser software tool;
Figure 7 is an example of an embodiment of a user interface displayed with a detailed outline; Figure 8 is an example of an embodiment of a user interface of the software browsing tool using an auto-scrolling method for presenting multimedia data; Figure 9 is an example of an embodiment of a user interface of a software browsing tool that presents multimedia data as a stacked slide show; Figure 10 is an example of an embodiment of a user interface displayed with the browser tool using the flyby effect; Figure 11 is an example of an embodiment of a layout technique that may be used with the flyby effect in presenting a group of multimedia data; Figure 12A is an example of an embodiment of a top-down view of multimedia data presented using the flyby effect and layout technique of Figure 1 1 ; Figure 12B is an example of an embodiment of a layout technique that may be used in a multimedia data presentation; Figure 13 is an example of an embodiment of a schematic for a technique for presenting multimedia data as in Figure 12B; Figure 14 is an example of an embodiment of user controls as included in various user interfaces for the browsing tool; Figure 15 is an example of an embodiment of data flow in a system for creating and using indices for multimedia data in conjunction with the browsing tool;
Figure 16 is a flowchart of method steps of one embodiment for generating and using indices with a synchronized multimedia presentation; Figure 17 is an example of an embodiment of a database schema record for indices created and used in connection with multimedia presentation to be browsed using the browsing tool;
Figure 18 is a more detailed example of the database schema record of Figure 17; Figure 19 is a flowchart of steps of a method for selecting a multimedia presentation using the browsing tool and multimedia presentation indices from the database; Figure 20 is an example of an embodiment of a data structure used to associate categories with one or more multimedia presentation items;
Figure 21 is an example of an embodiment of a data structure used to associate indices with each multimedia presentation item; Figure 22 is an example of an embodiment of a user interface display of video titles using the browsing tool; Figure 23 is an example of an example of an embodiment of a screen displayed when a selection of a movie title has been made using the browsing tool; Figure 24 is an example of an embodiment of a user interface display of a multimedia presentation using the browsing tool; Figure 25 is an example of an embodiment of a screen displayed when a selection of a multimedia presentation has been made using the browsing tool; Figure 26 is an example of an embodiment of a user interface display of multimedia shopping titles displayed using the browsing tool; Figure 27 is an example of an embodiment of a screen displayed when a selection of a multimedia shopping title has been made using the browsing tool; Figure 28 is an example of an embodiment of a system that includes a journaling application for recording and playing back a multimedia presentation; Figure 29 is an example of an embodiment of a user interface display of a journaling application; Figure 30 is an example of a flowchart of an embodiment of steps of a method for capturing and viewing multimedia information;
Figure 31 is an example of an embodiment of a time line as may be used to represent the time coordination of events in a multimedia presentation; and Figure 32 is a flowchart of method steps of an embodiment for synchronizing interrelated components of a multimedia presentation.
Disclosure of Invention
In accordance with one aspect of the invention is a method and computer program product for selecing a multimedia presentation. A subset of multimedia presentations are determined in accordance with predetermined criteria. One or more multimedia data items are determined. Each of the one or more multimedia data items is an index into a corresponding one of the subset of multimedia presentations. The one or more multimedia data items are viewed using a browser. Direction and speed of the one or more multimedia data items viewed is controlled. A first of the one or more multimedia data items is selected. Control is transferred to machine executable code associated with a first multimedia presentation having the first multimedia data item as an index.
In accordance with another aspect of the invention is a method and apparatus for selecting a multimedia presentation. One or more indices are determined for a multimedia presentation. The multimedia presentation has a first media stream and a second media stream. A portion of the first media stream is used as an index into a portion of said second media stream. Information about the indices is stored in a database. Multimedia objects are displayed in accordance with the information about the indices in the database using a browser. Direction and speed of the multimedia objects displayed are controlled. A first of the multimedia objects is selected. A software program is invoked for presenting the multimedia presentation having the first multimedia object as an index.
In accordance with yet another aspect of the invention is a method and apparatus for selecting a multimedia presentation. A subset of multimedia presentation is determined in accordance with predetermined criteria. One or more multimedia data items is determined. Each of the one or more multimedia data items is an index into a corresponding one of the subset of multimedia presentations.
One or more multimedia data items is viewed using a browser to select a multimedia presentation. Direction and speed of the one or more multimedia data items viewed is controlled. A first of the one or more multimedia data items associated with the multimedia presentation is selected.
Best Mode for Carrying Out the Invention
Referring now to Figure 1, shown is an embodiment of a system. A computer system 10 is shown to include a network 12 by which computer nodes 14- 18 may communicate with each other. In the computer system 10 of Figure 1 , one of the computer nodes, such as node 15, acts as a client node upon which a user executes a client browser 16. The client node 15 communicates with one or more other nodes in the system 10 in which the one or more other nodes in the computer system 10, such as node 14, may act as a server node providing certain functions or responding to certain requests made by the client browser executing in the client node 15.
It should be noted that the system of Figure 1 may include one or more computer nodes and the nodes, if there are more than one, may communicate through any type of communications medium, such as a network, a hardwire connection, or other means of communication known to those skilled in the art. It should also be noted that each of the computer nodes in the system 10 may be any type of computer processor, such as a commercially available personal computer, or a larger scale multi-user computer system.
Referring now to Figure 2, shown is an embodiment of various software components that may be included in one or more nodes of the computer system 10 of Figure 1. Generally, the software components in Figure 2 may be used to produce the files needed to build a software browsing tool. Additionally, some of the components included in Figure 2 are files that may be used by the software browsing tool once complete. Generally, the software browsing tool that will be described in paragraphs that follow is a tool providing overviews of dynamically specified data for the purpose of, for example, previewing or assessing data. Generally, the tool presents an organizational structure by which a user may view various data components and objects, such as those associated with image -based and/or multimedia data presentations.
Included in Figure 2 is a hierarchical data file 20, user interface components 22, multimedia files 24, and a translation tool 26. The hierarchical data file 20 generally includes a description of the various data components to be presented and viewed by a user in some type of hierarchical organization. In one embodiment, the hierarchical data file may include a list of the various URLs specifying images and/or other multimedia data files to be presented to a user. Additionally, various URLs are grouped together to represent the grouping hierarchy by which these components are to be presented. For example, the hierarchical data file may include a nesting structure nesting various groups of one or more of the URL identified files presented to the user. The user interface components 22 generally include software used to produce the user interface displays with the browsing software tool. As will be described in paragraphs that follow, user interface components may include, for example, JavaScript code and data, and dynamic HTML files. Multimedia data files 24 generally include multimedia data to be presented to the user in conjunction with the user interface. Generally, the multimedia data files include that data which is organized and represented by the hierarchical data file. It should also be noted that other types of data files, such as multimedia data files including audio files may also be included in other embodiments. The translation tool 26 is generally used to produce other pieces of software which are used in displaying the user interface and associated data files. In one embodiment, the translation tool 26 includes Java code to produce files which will be described in conjunction with Figure 3.
It should generally be noted that other embodiments may include additional software components from those described in Figure 2.
Referring now to Figure 2A, shown is an example of an embodiment of a hierarchical data file as included in Figure 2. The hierarchical data file includes data 150 that may be grouped together as previously set forth in conjunction with Figure 2. In this embodiment, the hierarchical data file, for each grouping of one or more multimedia files, includes NAME/TITLE INFORMATION, one or more Related
URLs, and one or more Presentation URLs. Generally, the NAME/TITLE INFORMATION is a text description of the multimedia files that are the Presentation URLs. The NAME/TITLE INFORMATION may be included, as in a menu area, as will be described in following paragraphs. The Presentation URLs are one or more multimedia files presented to the user in a presentation area, for example, of a user interface. The one or more Related URLs may annotate or further describe the one or more Presentation URLs. For example, the Presentation URLs may identify images of various types of cars by a manufacturer. One of the Related URLs may identify a website of the manufacturer. The NAME/TITLE INFORMATION may be a text description identifying the manufacturer and type of cars presented. Another one of the Presentation URLs may be a video clip of a commercial for the types of cars presented. Yet another Presentation URL may identify an audio file of a recorded jingle or song by the manufacturer as used, for example, in a commercial. Referring now to Figure 3, shown is an example of an embodiment of the various data files used and produced by the translation tool 32 to display the user interface and provide the software browsing tool. The hierarchical data file 30 is used as input to the translation tool 32. The translation tool 32 may produce a variety of software output files. Generally, the software output files are included in the software browsing tool used to present data. The actual output produced by the translation tool is in accordance with the actual implementation of the various components of the software browsing tool. For example, in one preferred embodiment, the translation tool produces the VRML code 34. Generally, VRML code 34 is code written in a language called Virtual Reality Modeling Language. As known to those skilled in the art, a VRML player which interprets or executes the VRML code 34 may be included in browsers such as the Netscape 4.0 or the Internet
Explorer 4.0 browser. The VRML player, for example, is one component required to execute the VRML code which implements one of the components of the browsing software which will be described in paragraphs that follow.
The translation tool 32 may also produce JavaScript code and data used with, for example, dynamic HTML files 36. Generally, if an implementation or embodiment includes the use of HTML files or VRML code 34, the translation tool 32 may generate JavaScript code and data.
The translation tool 32 may also produce Java Code 38. Java Code 38 may be used as an alternative to writing various components of the software browsing tool using the dynamic HTML files.
It should generally be noted that the software components of Figure 2 as well as additional software components included in other embodiments may be included on one machine, or on any combination of different nodes in the system of Figure 1. It should also be noted that various embodiments may include various restrictions in accordance with the particular versions of software, for example, used in a particular embodiment. For example, if the translation tool 32 of Figure 3 includes Java code which is run using a version 1.1 Java applet with the default security model, it should be noted that this version of Java is only able to load components from the same machine that it was loaded from due to the security model of that version. Thus, various components described in Figure 2 must reside on the same computer node of Figure 1 , such as the Java code used for the translation 32, the multimedia files 24, such as sound files, the user interface components, and the hierarchical data file 20.
In contrast, if this default is changed or another version of Java is used to execute the Java code included in the translation tool 32, this restriction may be removed. Thus, the various components described may be located on other nodes of the computer system 10 of Figure 1. It should also be noted that when different software components as included in Figure 2 are located on different nodes in the computer system 10 of Figure 1, different transferring mechanisms may be used to transfer files. For example, in a network environment with the Internet protocol, HTTP may be used. Additionally, FTP may be used to fetch files.
Referring now to Figure 4, shown is a flowchart of an example of an embodiment of the method steps of what happens at run time when a client browser 16 issues a request for the software browsing tool. At step 21, the client makes a request. For example, referring back to Figure 1, the client browser 16 on node 15 may issue a request using network 12 to communicate to another server node, such as node 14. At step 23, the user interface components located on a node, such as server node 14, are loaded into and executed on the client system by the client browser 16. At step 25, execution of the user interface components causes the client browser 16 to make a request to the server node 14 for the hierarchical data file. Additionally, execution of the user interface components causes the translation tool to execute producing the various software components of the software browsing tool and subsequently sending these software components produced to the client node 15. At step 27, these translated files produced by the translation tool 32 are executed by the client browser on the client system 15 causing images and other multimedia files to be loaded in accordance with a particular embodiment. It should be noted that depending upon where the various software components of Figure 2 are located, various communication requests may need to be issued by the server node, for example, to obtain the necessary files to produce the software components. Additionally, as noted at step 27, the precise images and other multimedia files loaded are in accordance with each particular embodiment. For example, use of the
VRML player causes all images to be loaded by default. Other software tools may have other defaults as well as variants which may allow image files and other multimedia data files to be loaded on demand. It should also be noted that in an embodiment where all of the files in processing occurs on the client system, no additional communication may be needed since the client and server node are the same. Referring now to Figure 5, shown is a block diagram of one example of an embodiment of a software browsing tool. Generally, the software browsing tool of Figure 5 includes driver commands and data 40 and various software components 46-50. The software components may vary with each particular embodiment of the software browsing tool and the functions provided, as well as the tools used to implement the software browsing tool. In this particular embodiment, the driver program and data 40 includes a driver program or script 42 and current focus data 44. Generally, the driver program or script 42 is a set of commands or statements which drive or control the software browsing tool. The current focus data 44 is generally a common shared data area describing the focus of the data presented to the user.
When the user is viewing data, the current focus of the data being presented may be changed as the user cycles through and views the various multimedia data items. As this happens, the focus changes as the user examines different images or other multimedia data. When there is a change in the focus, the focus data 44 is updated by a signaling handler as specified in the driver program or script 42. A mechanism exists in the driver program or script providing a control for updating the focus data 44 and additionally signaling all of the components as to the change in the current focus. The various components then update their local copy of the focus data and redisplay their various portions of the user interface. In this particular embodiment, for example, the driver/program script 42 is a JavaScript and the current focus data is represented as a JavaScript object. All of the URLs which are presented to the user representing multimedia data are represented as JavaScript objects with the grouping structure or hierarchy described in the hierarchical data file 20. The driver program or script 42 may be one of the JavaScript code and data components produced by the process of Figure 3 by the translation tool 32. The VRML runtime component 46 may include VRML code 34 as produced by the translation tool 32 of Figure 3.
It should be noted that other embodiments may implement the current focus data as other language entities or data objects in accordance with the implementation language chosen. The various other components, such as the control bar component 48 and the sound player component 50, may be implemented using a variety of different coding languages. In one embodiment, Java code, as may be produced by the translation tool 32, may be the language of implementation for these components. In another embodiment, one or more of these components 48-50 may be implemented using
JavaScript code and data with dynamic HTML. The precise control mechanism used by the driver program/script 42 may also vary with implementation and embodiment. In this particular embodiment, the Java scripting language provides a control mechanism by which the various components are updated when the user focus changes. It should be noted that each of the components 46-50 of Figure 5 for the browsing software tool are generally independent, pluggable pieces that may be removed or added independent of the other components. Generally, as will be described in paragraphs that follow, each component maps to a particular component of the user interface displayed and is responsible, for example, for updating and performing other functions within that particular portion of the user interface.
It should also be noted that the hierarchical data file 30 of Figure 3 may be produced manually or in an automated fashion, such as by a software tool that may be implemented using one of many different languages known to those skilled in the art. For example, various directory structures and subdirectories may be used to group files in accordance with the hierarchy description to be included in the hierarchical data file. A software program written in a commercially written programming language, for example, may use system functions to inquire as to the contents of the various directories and subdirectories and accordingly, produce a hierarchical data file in accordance with the directory and subdirectory structure. Other embodiments may include different ways of producing the hierarchical data file in an automated fashion in accordance with a particular system of Figure 1.
It should also be noted that the various images or other multimedia data files which will be used in presenting information to the user by the software browsing tool of Figure 5, may be loaded prior to runtime, such as by caching data in memory on the server or other node as part of a preprocessing step. Referring now to Figure 6, shown in an example of an embodiment of a user interface as displayed by the software browsing tool. Generally, Figure 6 shows an example of the browsing tool used in the application of the domain of Internet shopping. The user interface 60 includes a menu or outline portion 62, a presentation area 64, and user control buttons 66. Generally, the outline or menu portion 62 displays various topics or text descriptions which accompany the items, images, or other multimedia data being presented in the presentation area 64. The user may control the speed or direction in which the images or other multmedia data which correspond to various items listed in the menu area 62 are displayed to the user. For example, as will be described in paragraphs that follow, control display 66 includes buttons which provide for backward and forward review of the multimedia data presented in the presentation area 64. Accordingly, the active or highlighted portion, for example, of the outline or menu area 62 is synchronized with the presentation in accordance with the user positioning of the controls. If the user, for example, wants to preview multimedia data items that were already presented, the rewind button from the user control area 66 would be used also causing the menu portion 62 to have its active item updated in accordance with the multimedia data items presented in the presentation area 64.
In the menu or outline area 62, the active or current focus is denoted by highlighting, for example, a particular item in the menu portion 62. In one embodiment, as shown in Figure 6, the current outline or menu item has an arrow 63 next to it in the upper left hand corner. As the user is presented with multimedia data in different areas in accordance with different portions of the outline, the arrow in the left hand corner advances accordingly in the menu portion 62. Other techniques may be used to indicate an active or current focus on the outline area 62.
Other ways of indicating an active or current menu item in the area 62 include highlighting a particular area on the menu, or showing a color contrast from the rest of the menu items.
With regard to the user interface displayed in Figure 6, the content of the multimedia data presented has to do with promotional advertisements generally associated with online or virtual storage such as those available through the Internet that had been categorized by product type as well as occasion. For example, suppose the user has selected the category "Thanksgiving" through the button bar area 65. In this particular embodiment, the system generates the browser tool for the set of stores that have a promotional offering in the category selected by the option 65. As the user places the cursor over the arrow controls at the bottom of the main presentation area 66, images and/or other multimedia data representing the promotional content are presented one after the other in a semi-random spatial layout in the presentation area 64. Transitions between stores clear the presentation area. Additionally, the store whose promotional items are currently being presented in the presentation area 64 are indicated in the listing at the left by the appearance of a red arrow, as indicated in the outline or menu portion 62.
The user may control the speed of the presentation of the images and other multimedia data by choosing among the arrow controls 66. In this particular embodiment, the greater distance the cursor is placed from the center of the control area, the greater the speed in a particular direction. For example, two sets of arrows in the control area 66 go to the right indicating forwarding of the multimedia data presentation, and two sets of display arrows go to the left indicating reversing the multimedia data presentation. The farther to the left the cursor is placed over the arrows, the faster in reverse the multimedia data in the presentation area is displayed. This also parallels the fast forwarding controls as indicated by the arrows to the right in the control area 66.
It should generally be noted that the multimedia data presented in the presentation area 64 of Figure 6 and in other user interfaces which will be described in paragraphs that follow may be presented in a variety of different presentation arrangements. In this particular embodiment, the multimedia data is presented in a stacking arrangement, as will be described in more detail below.
Referring now to Figure 7, shown is a user interface of the browser software tool with a variant including more detail in the text outline portion 72. Figure 7 includes a presentation area 76, user control 74, and an outline or menu portion 72. These areas are similar to those as previously described in conjunction with Figure 6 user interface 60. This variant of the browser is also for the shopping domain in which the listing of the stores in the left hand portion includes store titles as well as the names of the products being promoted. In Figure 6, the user interface 60 included only stores in the outline or menu area. In this embodiment of the user interface of Figure 7, the outline or menu area 72 includes a store name, such as Disney™ 71 , as well as the product being promoted such as Thanksgiving
Pooh™73. In this variant, products are shown one at a time in order to make synchronization with the presentaton area 76 and the menu item 72 straight forward with the detailed hierarchical text outline listing. Thus, at execution time, the browser variant 70 of Figure 7 may present one image or other multimedia data element at a time. This is in contrast to the user interface 60 of Figure 6, in which multiple multimedia data objects or elements are presented which may overlap or be superimposed upon one another.
It should generally be noted that the multimedia data presented in the presentation areas 76, 64, and other user interfaces which will be described in paragraphs that follow, may be "dragged and dropped" for performing other user functions. For example, the user may drag and drop an item from the presentation area to place in a virtual shopping basket for items which the user may wish to purchase at the end of their viewing or browsing session. Additionally, by selecting one of the items in the presentation area, such as by a double click with the left hand mouse button, a user may be connected, as by a hyper link, to a different Internet website. Selection may also have other meanings depending on the application of the viewing or browsing session. For example, if the user is selecting or viewing movies or TV channels, selecting a particular item in the presentation area may mean to select a movie and view a portion of a movie trailer, for example. If it were a TV channel selection, an image selection from the television show on each particular channel may be presented in the presentation area 76. The user may select a TV channel and the particular TV program for current viewing.
Additionally, when multimedia data is presented to a user, various multimedia files may also be executed, for example, while an image is being viewed. For example, a sound file in addition to an image file may be played to a user when viewing images, for example, of a particular movie. It should also be noted that in the menu areas 72 and 62 a user may jump to a particular section in the sequence by selecting an item from the menu For example, if the user's focus or current point is at the top item as in Figure 6, a user may advance to the end of the menu item and accordingly cause presentation of the multimedia data associated with that item by selecting the last item from the menu area This is a shortcut method by which a user may quickly advance to a particular area in the menu to look at and view the images and other multimedia data associated with that particular item in the menu selection.
The user interface of Figure 6 may generally be described as a slide show presentation by which the user may begin a multimedia presentation associated with the first item in the menu area 62. An embodiment of this may automatically advance at a set speed from the beginning to the end of the menu and accordingly present multimedia data in the presentation area 64 at a given pace. The user may control the presentation of the multimedia data by using the control area 66 This is similar to the way in which a movie, for example, may be viewed in which the play button causes the images of the movie to be presented at a particular speed. However, a user may rewind or fast forward to a particular section using the controls This is a method of auto-scrolling through multimedia data associated with the menu selection area 62 The multimedia presentation of Figure 7 generally includes a hierarchical level of menus displayed in the area 72 In this particular embodiment, there is a nesting of several levels of menu items Generally, the outermost level menu is denoted as level 1 and the level numbers increase sequentially to 3 and the like as the nesting gets deeper In this particular display 72, while an innermost level is displayed, the outermost levels also stay constant on the menu For example, if the nesting level was level 3, and the menu items spanned more than one screenful, while the additional items from the level 3 menu were displayed, the level 1 and level 2 items would appear constant on the screen for the user so that the user is provided with information indicating where the current focus is for the menu items being displayed In other words, the user is provided with context feedback information describing where at any particular time multimedia data m the presentation area is located relative to the menu items 72. The presentation area 76 of Figure 7 may present images and other multimedia data using a variety of techniques. Multimedia data may be presented in a temporal fashion, one multimedia data item at a time with subsequent data items replacing or overlaying previous multimedia data items, in synchronization with a sound file and the more detailed menu 72, for example. Multimedia data may also be presented in a rolling or scrolling fashion in an upward or downward direction. This technique is further described below.
Referring now to Figure 8, shown is an example of a user interface of the software browsing tool which presents the auto-scrolling method for presenting multimedia data. The user interface 75 includes user control 77, a presentation area 75, with multimedia data presented, such as multimedia data items 202 and 204, and a menu or outline area 78. In this variant of the user interface of the browsing tool, the user scrolls through a series of multimedia data items which are predetermined in a format layout out such as in a top to bottom fashion. The motion of controlling or scrolling through the multimedia data using control 77 scrolls along in the x-y plane as indicated by the arrow on the right hand side for forward and reverse direction of presenting the multimedia data items. A portion of multimedia data items 202 and 204 are shown as extending beyond the user interface screen 75 for the purposes of demonstration only. Generally, if the user were to, for example, rewind or reverse through the user controls 77, a portion of the multimedia data 202 would be presented which is off the screen for the current point of view. Similarly, if the user were to slightly fast forward using the control 77, a portion of the multimedia data 204 which is shown as off the screen currently would be included on the presentation area 52 displacing a portion of the other multimedia data items, such as 202. Thus, the user has a point of view or focus which is the size of the presentation area 52 which scrolls downward through a prearranged presentation of multimedia data in a layout fashion. By using the user controls 77, a user may view portions of the multimedia data in the presentation area 52. As described before in conjunction with other figures, the user may advance to a different position on the scroll by selecting a particular menu item. This is in addition to using the user control 77 to rewind and fast forward to the multimedia data contained in the scroll.
It should be noted that "Level- 1" and "Level-2" of menu area 78 may be "active" in that they may be selected and used to modify the current focus or point of view of the user with regard to the data presented in the presentation area 52.
Referring now to Figure 9, shown is an example of an embodiment of the user interface for the browser tool which presents multimedia data in the presentation area 52 in a stacked slide show fashion. Generally, note that multimedia data items which may be presented in the presentation area 52 are slightly stacked upon one another for viewing at a predetermined time interval. The user may control with the user controls 77 how fast or slow the multimedia data items are stacked upon one another in the presentation area 52. It should generally be noted that the multimedia data may also appear in an unstacked fashion in the presentation area 52 such that the multimedia data items do not overlap each other. This is yet another variant of the browser presentation area for presenting multimedia data in accordance with a menu item 78 with user control over the presentation of the multimedia data 77.
Recall that the scrolling effect had the images and/or other multimedia data presented in a predetermined top to bottom fashion with the user interface presentation area acting as a window clipping or viewing a portion of this predetermined area. The stacked slide show is similar to the appearance of having a deck of cards where each multimedia data item corresponds to a card in the deck in which the cards are placed one on top of each other in the presentation area. Fast forwarding or rewinding through a stack slide show of multimedia data results in cards being removed or placed on top of the deck. Similarly, it would result in multimedia data items being stacked or layed on top of one another or removed from the stack.
Referring now to Figure 10. shown is an example of a user interface of the browser variant in which the presentation area or the display area incorporates a three-dimensional fly through or flyby effect as opposed to the previously described scrolling effect in various slide show effects. The user may control the presentation in the presentation area 86 using the control arrows 84 The user may select various categories of multimedia data items to be presented using the buttons indicated in the selection area 82 of the user interface display 80 Additionally, as previously described in conjunction with other presentation areas, a user may select a particular multimedia data item from the presentation area 86 which causes the user to jump to the page where the multimedia data item is from In other words, selecting an item from the presentation area such as with the mouse, a user may use the hyperlink Internet connection by which they are connected to the page on the Internet upon which an image or other multimedia data item may be located Generally, the position of the user's mouse cursor, for example on the arrow 84, determines the speed of the presentation which essentially has an infinite range from 0 to the maximum that the display device is capable of achieving Audio feedback of the current speed may also be included Generally, the three dimensional flyby effect, which will be described in more detail in paragraphs that follow, involve the user being presented with images or other multimedia data as if they are flying by the user Generally, multimedia data may include movie or video clips, sound files, or 3- D models, and the like Each of the multimedia data presented is subject to a 3-D flyby effect with regard to the type of multimedia data For example, if a multimedia data file presented is a sound file subject to the 3-D flyby effect, the intensity of the sound is varied in accordance with proximity to the user
Referring now to Figure 1 1 , shown is an example of an embodiment of a frontal point of view of multimedia data presented using the flyby effect of Figure 10 The presentation arrangement 182 is one technique that may be used to present images in the presentation area 86 of Figure 10 Data items 160-170 are presented in as a group of data Data items are presented one at a time in a counter-clockwise direction in a circular arrangement The center of each data item 160-170 is placed on the circumference of the circle 184 as it appears in the presentation area 86 It should be noted that circle 184 does not actually appear in the presentation area but is displayed in Figure 1 1 to indicate placement of the data items 160-170 Each of the data items 160-170 are placed equidistant apart such that each of the angles 172-
182 are the same This technique may be generalized for any number of data items such that they are placed equidistant apart in a circular arrangement in a counterclockwise order.
Subsequent groups of data items are presented in an arrangement with a random angular shift in the counterclockwise direction. In other words, in subsequent data item groups presented, the position of the first data item 160 is shifted by a random amount in the counterclockwise direction. Similarly, the placement of subsequently presented data items in the same group are positioned with the same angular shift.
Referring now to Figure 12A, shown is a view of the multimedia data of Figure 1 1 looking down with the 3D flyby effect being used to present the multimedia data to the user. Each of the data item groups 122-128 represents a group of data items presented similar to those in the group 182 of Figure 1 1. In other words, each of the groups 122-128 is a "slice" or page of data items presented using the circular layout technique described in conjunction with Figure 1 1. Generally, the procedure for laying out multimedia data in a 3D flyby effect includes defining a certain area and placing the multimedia data for presentation within that defined area. A user's view or perspective at a particular point in time, such as a snapshot of the multimedia data in Figure 1 1 is depicted as having the user move along the arrow indicated in the left-hand side of Figure 12A on the x-y axis 69. Thus, at any particular point in time, a user's point of view or snapshot of the multimedia data, such as those presented in Figure 10 in the presentation area 86 represent a viewpoint of the multimedia taken from user perspective. For example, as the user fast forwards or proceeds through the presenting of the multimedia data items moving towards the screen or towards the data item 122, the data items 124, 126, and 128 will appear to fly past the user and the data item 122 will appear to come closer to the user. Similarly, when rewinding or moving out from the direction from data item 122 towards 128, the data items will appear to fly by the user in reverse order with a perspective or focal point into the screen.
Generally, the data items presented have a perspective such that the farther the multimedia data items are in the fast forward direction, the farther away they appear in the presentation to the user and appear to go to a focal point into the screen. Similarly, as multimedia data items are viewed, they pass by the user on the presentation area 67 giving the user the impression that the multimedia data items are moving past where the user is out of the visual path. Thus, when a user rewinds the multimedia data items in the presentation area, the multimedia data items are presented or refreshed in reverse order. This gives the appearance as if the data items are coming from behind the user to the current focus on the screen and, as the rewinding continues, the multimedia data items move into the screen such as at a distant point.
The top or downward view of the user perspective in Figure 12A generally includes a defined area 67 where groups of multimedia data items 122-128 may be positioned. The user's perspective or focus moves along the x-y axis, as indicated by the arrow 69. At any point in time, the user's view point may be associated with a point along the x-y axis 69. The multimedia data items are presented to the user from the user's perspective at a particular point along the x-y axis denoted by the arrow 69 as if the user were sitting in front of the defined area 67 with all of the multimedia data items coming towards the user or away from the user, as controlled by the control area 66 of Figure 6. For example, if the user were sitting in front of the screen and the controls indicated the position with regard to the user's perspective with a focus at 208, the multimedia data item associated with 128 appears closest to the user with groups of multimedia data items 122-126 giving the perspective of being farther away. In contrast, if the controls indicate that the user perspective is at position 206, no multimedia data items are presented on the screen and this may give the user the perspective that all of the multimedia data items have already been viewed. Referring now to Figure 12B, shown is an example of an embodiment of another technique for presenting data in a presentation area 120. Generally, the presentation area 120 may be included in one of the user interface displays such as 80. What will be described in conjunction with the next two figure is a technique for presenting multimedia data in a group layout in which there is a parent and child relationship between various multimedia data items to be presented on the presentation area 120. In this particular embodiment, there is a group of data items 190-198, with the parent 198 being located in the center of the presentation area and the children 190-196 being located surrounding the parent. In this particular embodiment of a group, the parent is presented in the center of the group with children being placed in four quadrants beginning with the upper left relative to the center of the group which includes the parent node. It should generally be noted that if there were only three children for a particular group, the children would be placed in order beginning with the upper left quadrant proceeding in a clockwise fashion until all the child data items have been presented. Similarly, it should be noted that if there are more than four data items, the fifth succeeding data item would overlap the first data item in the upper left quadrant and accordingly be placed in that quadrant with succeeding data items numbered 6, 7, 8, and so on overlapping successive quadrants. Using a slight variant of this technique, rather than successively present multimedia data in each of the four quadrants, one may view the upper left quadrant as a starting point and an end point and present multimedia data items in a clockwise circular fashion one on top of another in a slightly stacked fashion beginning and ending with the upper left quadrant and accordingly space the multimedia data in accordance with the number to be presented in a particular group.
The snapshot of multimedia data items presented in Figure 12B represents the multimedia data items which may appear to a user, for example, at a particular point in time.
Referring now to Figure 13, shown is an example of an embodiment of a layout technique for presenting the data items of Figure 12B. As previously described, the technique with regard to presenting a particular group of multimedia data items is a layout between a parent data item and its immediate children. For example, in presenting a video and various snapshots from that video in the presentation area 90, the video title may appear in the center 1 10 as the parent node. The children may be the still images or snapshots of various pictures from the movie appearing as still images 100, 102, 104, and 106 in Figure 13. The title of the group is placed in the center of the canvas or presentation area. The technique then rotates placement of succeeding images such as 100, 102, 104, and 106, in a clockwise fashion around four quadrants 92, 94, 96, and 98 beginning with the upper left quadrant 92. The title multimedia data or parent 1 10 always remains on top of the still images and/or other multimedia data or children while succeeding multimedia data items may overlap each other. Each of the quadrants itself is divided further into quadrants, and a corner of each multimedia data item is randomly placed within the outermost quadrant. That is, each of the multimedia data items shown in Figure 13 has a corner placed somewhere in one of the rectangles 92, 94, 96 or 98. Such a technique for presenting multimedia data items provides a predictable temporal sequence by providing a regular rotation that is further reinforced through a layering effect in that more recent data items are always in a layer above the less recent ones.
Also included is a fading effect through control such that only the most recent data item is fully opaque with others successively less so. The randomness and the overlapping allows the use of images and other multimedia data items of various sizes and aspect ratios. Referring back to Figure 1 1, the technique described in conjunction with Figure 13 is applied to the multimedia data items produced and presented in the presentation area 120. In effect, the multimedia data items presented in Figure 1 1 are those at different points and times for different groups as they fly by the user.
The examples just described present user interfaces with images and/or other multimedia data items that particularly fall into two general domains which are
Internet shopping, and video selection or program guides. However, the techniques and the user interfaces are widely applicable to any variety of data being presented associated with any particular domain or application in which the content may be structured hierarchically. Individual objects or multimedia data in the content may be rendered in a form of image based summary information. The techniques described are capable of using a number of different spatial layout techniques and transitions for rendering individual multimedia data items to be presented in a particular presentation area.
In addition to the actual presentation of the multimedia data items to the user in one of a variety of forms and techniques in the presentation area, the actual user controls as included in the control areas of the previously described user interfaces are also important.
One advantage of the foregoing technique is that each of the multimedia data items may be presented for a longer time period than using the prior art slide show technique improving recall and depth of processing of a particular set of objects when presented to a user. In other words, the combined temporal and spatial presentation of the foregoing technique provides coherence in associating content from the same grouping of objects with each other.
Referring now to Figure 14, shown is an example of a variety of user controls as included in the previously described user interfaces. The controls allow the user to make adjustments to the speed as well as the direction of the temporal presentation. The user controls 130 may be used to control both direction and speed in which the position of the mouse cursor, for example, within the arrow is an indicator of the speed. As the mouse arrow approaches the arrow tip, the speed increases up to the maximum that the display device may support. Visual feedback is indicated to the user by the level of color saturation within the arrow. In other words, the faster the images/and or other multimedia data items are presented, the closer to the arrow tip the mouse cursor is, and the darker or deeper the color of the display arrow. Additionally, instant on/off is achieved when the mouse is over the user control 130 or on the user control 130. In other words, the user does not need to select by depressing or clicking the mouse button in order to turn on or change the speed, or its direction. Simply by placing the mouse cursor over a particular portion of the arrow causes the speed to be altered or the direction to be changed. Cursor off of the arrow causes the speed to return to the predetermined rate, or stopping the presentation. Elements 136 and 138 are variants of the controls of the arrow depicted in 130. The slidebar 134 may be used also to slide to various points within the presentation of the multimedia data items in a continuous sequence. It should be noted that the function provided by the slidebar 134 is similar to the menu selection location as provided, for example, by the menu presented in the menu area 72 of Figure 7. Other embodiments of element 136 may have any number of additional arrow icon blocks in the forward and reverse direction offering a finer granularity of control. For example, the arrangement in 136 has one block of arrows for the forward and reverse direction. Another embodiment may include two or more blocks of arrows for each of the forward and reverse directions.
Generally, the different varieties of control buttons displayed as included in Figure 14 include a feature such that the speed and direction may be controlled by instant on off with a mouse cursor, for example, entering or leaving a display icon corresponding to the user control area. The user controls determine the speed and direction of a presentation in which the movement of the mouse cursor, for example, over a spatial area of the screen starts the presentation of a sequence of multimedia data items. Movement within a spatial area corresponding to the directional arrow determines the speed of the presentation. Subsequently, movement away from the spatial area corresponding to the user control arrow stops the presentation. Generally, the mouse over the arrow button area serves as a trigger for starting a temporal sequence and the mouse leaving this predetermined area causes the presentation of the multimedia data items to cease. This is in direct contrast to prior designs which require mouse clicks, for example, indicating users selection to start and stop display. With a rapid display of objects or images in the presentation area, users invariably may pass their intended stop location prior to physically depressing and selecting via the mouse control.
In the previously described user interfaces and browsing tool, the use of two or more synchronized displays provides a traversal in browsing of multimedia data items in a hierarchical structure. The presentation of the multimedia data items is generally controlled temporally with forward and backward and faster and slower controls for the purpose of content selection or assessment. The multimedia-based content of the individual items in this sequence may be rendered in a designated spatial container. Generally, multimedia data items are presented in a synchronized fashion with two or more components of the user interface, such as the menu selection or outline and the presentation area upon which multimedia data items are presented. This provides a technique for viewing or selecting a set of information object out of some larger collection. One technique was presented for grouping objects with various structure relations, such as the parent/child relation and the 3D flyby presentation.
The foregoing techniques are flexible in that they may be applied to a variety of domains of multimedia data and applications. Using combinations of the controls and the selection techniques, images and other multimedia data items may be presented for efficient selection based on content.
The foregoing techniques may be used in applications for browsing and selecting from multimedia data. One such system that includes the foregoing browsing techniques will now be described.
Content selection may be performed in connection with multimedia data from a variety of applications such as those using the Internet. The foregoing browsing techniques may be used in indexing into multimedia data using spatial and temporal aspects. In particular, what will be described is a synchronized spatial- temporal technique for browsing images for the assessment of content. Image- oriented data associated with interactive multimedia titles, for example, may be used to browse and index into different varieties of multimedia data associated with the titles. Due to the temporal nature of some forms of multimedia data, such as those involving audio and/or video data, it may be difficult to "drill down" or narrow a search to a subset of large volumes of data. The foregoing browsing techniques may be employed in a system that provides assistance in browsing and selecting from a subset of multimedia data.
Referring to Figure 15, shown is an example of the embodiment of data flow in a system for creating and using indices for multimedia information in connection with a browsing tool. The system 300 includes multimedia information 302 for which indices are created by an index extraction process 310 with an optional filtering process 312. Database records corresponding to, and representing, the multimedia indices are created 314 and stored in the index database 308. The browsing tool 316 uses information regarding the indices, as read from the index database 308, to allow the user to browse various multimedia indices. Using this browsing tool 316, a user selects a particular index and control is transferred to an application, such as 304 or 306, for presenting the corresponding multimedia information 302.
It should be noted that the spatial/temporal image browser 316 shown in Figure 15 in the system 300 may be an embodiment of the previously described browsing tool as described in conjunction with the computer system 10 of Figure 1.
Generally, as will be described, the browsing tool 316 displays various multimedia information in one of several display formats, such as those previously described, to aid a user in the selection and viewing of various portions of multimedia presentations as may be stored in the multimedia information 302. The user then selects a particular index corresponding to a particular multimedia presentation as may be included in the multimedia information 302. Once this selection is made, control may be passed to an application, 304 or 306, for displaying a multimedia presentation.
The multimedia data which may be extracted and used as an index for the various multimedia presentations includes any one of a variety of multimedia types, for example, such as image data, audio data, other video data, view graphs and other media streams.
Generally, the multimedia information 302 may include different types of multimedia presentations and information and may further include one or more varying streams of information combined together to make a single multimedia presentation. For example, as will be described, one multimedia presentation may include four media streams in connection with a presentation such as a speaker that performs the presentation with view graphs. The four different media may include, for example, view graphs, an audio stream, a video stream and speaker notes. A second type of multimedia information that may be included in the system 300 includes video key frames that index into a video stream and an associated audio stream. This is an example of a video clip having a corresponding audio track, for example, as when viewing movie trailers or clips. Such multimedia information and resentation may be used, for example, in connection with a video-on-demand application. A third type of multimedia information that may be included in the system 300 may be related to consumer online shopping, as may be included in an electronic commerce application providing product descriptions and information to consumers shopping over the Internet. This type of multimedia presentation may include several different types of media streams similar to the video-on-demand. For example, there may be a presentation or a video clip for a product demonstration as well as additional text and audio information, such as a jingle or a verbal description, placed in an audio file accompanying the product demonstration. Each of these items may be included in a different media stream which, when in combination, result in multimedia presentation for a particular product.
Referring now to Figure 16, shown is a flowchart of method steps of one embodiment for generating and using indices with synchronized multimedia presentations in the system 300 of Figure 15. Generally, what will be described in conjunction with Figure 16, flowchart 330 are method steps relating to the data flow and description previously set forth regarding the system 300. At step 332, for each multimedia data item or presentation included in multimedia information 302, one of the media types associated with the presentation is chosen as the indexing media. In one embodiment that includes multimedia presentations which are speaker presentations, using the previously described four media streams, view graphs may be used an index. A time stamp may be used to synchronize a particular point in the video, audio and speaker notes accompanying the corresponding view graph. Each view graph may be associated with a time stamp and a corresponding point found within each of the other media streams. In one embodiment, which includes the video or movie presentations, a video key frame may be used an index into a corresponding audio stream. In another embodiment that includes the online shopping multimedia presentations, a product image which represents a key frame in a video such as a product demo may be used as an index similar to the video key frame previously described with regard to a movie. The video key frame representing a product image is similarly used as an index into accompanying multimedia streams associated with a particular product being represented in a multimedia presentation.
At step 334, image-based indices are generated in accordance with the indexing media. It should be noted that at step 334, if other types of indices besides those which are image -based are used, these are also generated. It should be noted that at step 334 the image-based indices which are chosen are those which will be displayed in a particular format, for example, as by the browser tool 316 in one of the previously described display layouts to allow a user to select a particular multimedia presentation in accordance with the displayed multimedia indices. It should also be noted that different processes for generating the image-based indices at step 334 may vary in accordance with the particular media streams as well as the variety of software that may be available in each particular implementation. Example embodiments and processes for generating different types of indices will be described in more detail in paragraphs that follow. However, it should be noted that this should not be construed as a limitation as to techniques used to generate indices which are image-based used in conjunction with multimedia information 302.
It should also be noted that techniques providing for the automatic extraction of key frames from video streams are known to those skilled in the art. For example, the shot detection technique is an example of a well-known automatic extraction technique as may be implemented with the commercially available software Virage
VideoLogger software system.
It should also be noted that at step 334, selection of particular portions of a medium, such as selecting a portion of a video frame to be used as a key frame and an index, may be done in an automated fashion as well as a manual fashion, or in combination. For example, there are techniques which determine that a video frame may be a video key frame in accordance with the amount of change detected between two consecutive images in a video. This may be done in an automated fashion, for example, using software which determines this difference in accordance with the data type and structure of the video files. Another key frame selection technique that may be performed, such as using videos with regard to the multimedia shopping presentation, may be performed manually. A video key frame may be selected as one to be used as an index, for example, if that frame of the advertisement for a product includes all of the components of a product in clear view. This may be performed manually by a person viewing the advertisement or clip for a particular product. Regarding the multimedia presentation of the speaker, each view graph included may be used as an index medium.
At step 336, optional image filtering processing is performed to remove any extraneous images. Step 336, for example, may be performed in an automated or a manual fashion. A person may view all of the image indices associated with a particular multimedia presentation and determine that one should be removed for any one of a variety of reasons in accordance with predetermined criterion. For example, a threshold number of indices may be chosen and a decision may be made to remove the first and the last image index for a particular presentation. Alternatively, for example, an automated process may be performed upon the selected image indices associated with a particular multimedia presentation to remove one or more images that do not meet predetermined criteria. Such predetermined criteria, for example, may include images of a particular color, including certain objects, or other criteria. One or more indices may be removed or "filtered out" in connection with, for example, heuristics related to the amount and/or number of colors, such as concentration and contrast related to image coloring, degree of similarilty or difference between two images, and temporal proximity.
At step 338, a structured index object is created that may be stored in the database 308. As will be described in conjunction with other figures, at step 338, one or more records are generated in accordance with the various indices. At step 340, the multimedia presentation as may be included in multimedia information 302 is updated if needed. For example, at step 340, a time stamp associated with each view graph of a multimedia speaker presentation may be determined. It may then be necessary to accordingly update one or more of the associated multimedia streams also associated with that presentation to synchronize the multiple media streams of the presentation. This enables, for example, a synchronization point in each of the media streams for a particular view graph. Similarly, this update step may be performed for other types of multimedia presentation, such as the video and audio streams associated with the video-on-demand selection. Step 340 may be performed prior to this point in processing as included in the flowchart 330.
However, since there is an optional filtering step 336, it may be beneficial to perform this step subsequent to performing any optional filtering so as not to unnecessarily update other media streams in the multimedia information 320 for indices which may be removed by the optional filtering step 336. Each technique for updating the different streams in accordance with, for example, the indices selected, may vary with each particular embodiment and implementation. Particular examples will be described in paragraphs that follow.
At step 340, the index database 308 may be updated to include the information for the indices created or determined for each particular multimedia presentation. Referring back to Figure 15 with system 300, what has been described in conjunction with Figure 16, flowchart 330 is the process by which indices are extracted from the multimedia information 302 and information may be stored in the database 308 corresponding to the various multimedia indices. As previously described, the multimedia information 302 for which multimedia indices are determined may include, for example, an entertainment movie, a seminar style presentation, and an online shopping presentation, as well as other types of multimedia presentations in accordance with applications.
It should be noted that, as described in other sections included herein, the indices may be presented for viewing in accordance with a predetermined structure or format. For example, an HTML file may be dynamically generating using information from the index database for each multimedia presentation to be presented in accordance with a user selection, such as seminars or speaker presentations may be displayed in accordance with a predefined viewing format for a particular speaker selected by a user. The format or viewing structure of the various titles may be predetermined. However, the HTML file may be dynamically generated in accordance with the particular user selection.
Referring now to Figure 17, shown is an example of an embodiment of a database schema record for indices created and used in connection with multimedia presentations to be browsed using the browsing tool. The record 350 is an example of a database schema record that may be stored, for example, in the index database 308 of the system 300 of Figure 15. The record 350 in this embodiment includes three fields. It should be noted that although what will be described are three fields, other embodiments may include records that include a varying number of fields with other additional information in accordance with each particular implementation. The record 350 includes a pointer to the image to be displayed by the browsing tool 316 in the field 350a. Also included in the record field 350b is an index or time stamp reference into the multimedia presentation. Filed 350c specifies a software application to which control is to be transferred. In this embodiment, field 350c includes a URL with a file extension. The URL may identify a multimedia presentation and the file extension may identify a software application that is invoked to present the multimedia presentation. It should be noted that other techniques may be used to both specify the software application and the multimedia presentation in accordance with the functionality available with each particular embodiment.
Field 350a, which is a pointer to the image to be displayed by the browsing tool 316, may be a file identifier or handle identifying an associated image, such as a file identifier of a JPEG or .GIF image file. The information 350a is associated with or represents an address of an image that is displayed, for example, such as an image included in the display format 182 of Figure 11 such as images 160 or 162. Field
350b identifies an associated temporal index or reference into the multimedia presentation. In particular, in this embodiment, the index or time stamp represents a reference with regard to the image associated with 350a. For example, with the video-on-demand application, the field 350a may be a pointer to an image which is a portion of a video in which the video key frame associated with field 350a occurs at a particular point in time with respect to a video clip. The point in time is referenced by field 350b. This index may be, for example, a temporal offset into a video with regard to playing time from the beginning of the clip. Field 350c represents a software application which, if the associated image is selected at field 350a, is invoked to display the multimedia presentation. In other words, field 350c is associated with a software application to which control is transferred to present one of a variety of different multimedia presentations.
Generally, the record format 350 of Figure 17 includes information which is displayed by the browsing tool, such as the images associated with fields 350a. A user selects one of the associated images displayed, and accordingly, a software application is invoked through a control-transfer mechanism in accordance with each particular embodiment. The particular software application is identified in field 350c. The software application may be transferred control along with additional information such as the index time stamp reference into the multimedia presentation represented by field 350b.
Referring now to Figure 18, shown is a more detailed description of a particular database schema record previously described in conjunction with Figure 17. In this particular example 360, information may be included in the record in accordance with the previously described record structure for a particular multimedia presentation. Record includes an identifier of a file of a view graph miniature image in the first field, with the time stamp in the second field, and a software application identifier of a multimedia presentation player in the third field. In this particular example, the record 362 in the view graph miniature image file identifies a file, such as a URL or a file name, in accordance with a particular file system of the image that is to be displayed by the browsing tool 316. The second field includes a time stamp which represents the particular point in the presentation and in the associated multimedia streams which correspond to the particular image represented in the first field which has been chosen as an index into the presentation. The third field of record 362 is a presentation player URL. In this particular embodiment, a presentation player which has the name of the DejaView software application is identified by a particular file extension on a URL that may be included in this field. For example, a URL with the file extension ".DJVW" specifies that control is to be passed to a particular software application.
As will be described in following paragraphs, this application may be used to view a speaker presentation, for example, including viewgraphs, speaker notes, and an accompanying audio and video stream. In one embodiment, the third field may include a URL describing a particular presentation in which an association may be made between a software application and a particular file extension. For example, in this particular embodiment, the third field may specify a URL of a file with a file extension associated with a particular software application through a MIME type.
The browser may use an operating system supplied control mechanism for passing control from the browser to the specific software application. As known to those skilled in the art, the particular mechanism by which control may be passed is in accordance with each particular embodiment and implementation and operating system-supplied mechanism.
The first record 362 may be an index into a speaker presentation as previously described which includes view graphs as well as speaker notes and a video and audio stream corresponding to a speaker presenting the view graph. Record 364 may correspond to an index, for example, with a video key frame serving as an index into a movie. The first field of record 364 identifies an image which is a video key frame. Similar to record 362, the second field represents a time stamp which is a temporal spacing into the associated video at the position where the video key frame occurs. The third field identifies a pay-per-view application, such as by a ".PPV" file extension which result in invoking a pay-per-view application. Thus, images displayed with this associated index may be a series of indices into movies or videos displayed, for example, by the browser tool 316 when a user is selecting a movie for a pay-per-view application.
Record 366 may correspond to an index, for example, of a particular product of an e-commerce shopping application. The first field may be associated with a file that is a product image. The second field may represent a time stamp or index into a video, for example, corresponding to a product demonstration. The third field may include a URL to an e-commerce shopping application identified through a particular file extension and its corresponding MIME type.
It should be noted that the foregoing is just one example of an embodiment of a database schema and record format which corresponds to each of the various indices in accordance with each of the multimedia data types. These records may be created and stored in the index database 308. The database itself may be any one of a variety of types of database such as an object database or relational database that can be implemented with any one of a variety of commercially available packages. At this point in the process, information about the indices have been gathered and created and stored in database 308. What will now be described in the paragraphs that follow is how this information stored in the index database may be loaded and used by the browser 316.
Referring now to Figure 19, shown is the flowchart 370 of steps of one embodiment of a method for selecting a multimedia presentation using the browsing tool and multimedia presentation indices from the database 308. At step 372, a subset or reduced set of data has been obtained in accordance with user search criteria. Generally, step 372 has the purpose of reducing the candidate set of data for which information is to be displayed by the browser 316 to a reasonable subset, for example, that may be displayed in accordance with one of the previously described display layouts and techniques. This subset of information may be obtained using any one of a variety of existing search engines and techniques. For example, for each multimedia presentation included in the multimedia information 302, categories may be associated with each of these. Each of the multimedia presentations and corresponding information may be indexed and stored in some type of structure from which data may be stored and retrieved, such as a linked list or other data structure known to those skilled in the art. A user may perform an initial query in accordance with predetermined categories associated with each of the multimedia presentations. As a result of performing this query in accordance with one or more user selected categories, the names of various multimedia presentations may be the resultant set. In this instance, the results set may include handles or identifiers associated with each of the records such as 362, 364, and 366 stored in the database 308. It should be noted that the embodiment of Figures 17 and 18 in the database schema record include, for simplicity, only that information which is associated with each of the applications and images. Other information that may be included in the database records and not shown in the foregoing figures, for example, is the link by which a query may be performed of the database. In other words, if a user has determined that they would like to look at multimedia presentations or videos associated with comedy movies, additionally stored in the database 308 or elsewhere is a link identifying each of the indices associated with a video key frame of a comedy movie.
This information may be stored internally in the database 308 within each of the records, or separately in a different data structure or database schema. For example, if the category is a comedy movie, stored in the database 308 or elsewhere may be a list of database records associated given titles with predetermined categories. This may, for example, save time when performing a query of the database by having these records already pre-selected and predetermined for each of a particular category.
At step 374, a database query is performed to retrieve the records corresponding to indices for the subset of data to be displayed by the browsing tool 316. The information retrieved as a result of step 374, for example, may be the image URLs and other associated fields stored in each of the records. At step 376, a hierarchical data file is dynamically generated which represents these query results. For example, performed in the processing of step 376, the URLs may be organized in a format, for example, such as the hierarchical data file 150 of Figure 2A in accordance with a predetermined display format. In other words, the URLs may be arranged in a particular format or structure in accordance with the display or viewing format.
At step 378, the image indices are displayed using the browsing tool 316 in connection with the hierarchical data file. At step 380, one of the displayed image indices is selected by a user such as from a display area having one of the image display formats, for example, as shown in Figure 9 or Figure 1 1. As the result of selecting one of the indices associated with a multimedia presentation, control is transferred at step 382 to a software program or application which presents the multimedia presentation corresponding to the index selected. Referring back to Figure 15, and to the information included in the database schema as described in conjunction with Figures 17 and 18, the browser 316 may be used to display the indices. A user makes a selection. The image browser 316 uses the information from the third field in the record, such as record 364, to invoke an application such as the pay-per-view application. Associated with the displayed index is a particular multimedia presentation, for example, as stored in accordance with the subset of information at step 374. In other words, there is a link between the various indices and a multimedia presentation where, given a particular index, a multimedia presentation may be found. Similarly, given a multimedia presentation, the various indices in the database may also be found. This data structure may be stored, for example, in the form of lists predetermined as previously described. For a specific index selected, the multimedia information is obtained from the multimedia information block 302. It should be noted that the browser 316 may obtain additional information as needed from the multimedia information store 302 in order to transfer control to a software application. The software application to be invoked is represented in the third field, for example, of a database schema record
360 of Figure 18. Using an operating system or other implementation-specific transfer mechanism, control may be transferred to another application with the appropriate parameters and other information to invoke and display the multimedia presentation corresponding to the selected image index. It should be noted that the techniques available for performing a database query, for example, of the database 308, may be supplied with a commercially available database package that may be chosen in accordance with each particular embodiment. Application Programmer Interfaces (APIs) may be supplied to allow a user to perform a data query retrieving the appropriate information that is used by the image browser 316. Information as to which of the indices correspond to a particular multimedia presentation associated with the subset as used in step 374 may be stored in the database 308 or elsewhere.
Referring now to Figure 20, shown is an example of an embodiment of a data structure that may be used to map categories to movies. The data structure 420 represents a relation between a predetermined set of categories and associated multimedia presentation items that are movies. Similar data structures and others may be used to represent this and other types of multimedia information in accordance with each of the applications, such as 304 and 306. The data structure 420 includes a list of predetermined categories. Each category may be associated with one or more multimedia items. In this example, cat-1 422 is associated with movie 424 and movie 426. The categories may be a set of predetermined categories in which, using either automated and/or manual techniques, categories are mapped to multimedia items, such as movies 424 and 426. A user may be selecting a movie in connection with a pay-per-view application. The user may select a category of movies, such as comedy, that corresponds to one or more categories, such as cat-1
422. This reduces the selection space to a subset of movies associated with this category. In this instance, the user query may result in only a single category, cat-1 422, and only those movies associated with this category. It is from this list of movies that the user may select a single movie or video resulting in control being passed to a video-on-demand or pay-per-view application program.
Referring to Figure 21 , shown is an example of an embodiment of a second data structure associating a movie with one or more indices that may be represented in the database 308. Data structure 430 may be used in conjunction with data structure 420 to display one or more indices using the browsing tool 316 in connection with a user query. In this example, once one or more movies may have been selected using data structure 420. Data structure 430 may be used to obtain the one or more indices associated with each movie. It should be noted that other data structures besides 430 may be used to store and retrieve information as needed to display associated indices using the browsing tool 316. For example, a linked list rather than a table format may be used. Other variations known to those skilled in the art may be used with appropriate storage and retrieval techniques in accordance with each variation.
Additionally, the information represented in data structure 420 and 430 in combination relates to one embodiment in which the multimedia presentations of interest to the user are reduced in accordance with predetermined categories associated with each of the presentations. Other embodiments may use other techniques to reduce the number of multimedia presentations of interest to a user. Similarly, the data structures used in connection with these techniques may also vary. The storage location of these data structures may also vary in accordance with each embodiment and system configuration.
What will now be described in conjunction with Figures 22-27 are illustrations and examples of particular applications that may be invoked in accordance with different kinds of multimedia presentations whose indices are displayed by the image browser 316. Referring now to Figure 22, shown is an example of an embodiment of a user interface display 400 to display video titles using the browser tool 316. This may be an interface, for example, presented with a video-on-demand selection as a service presented with a choice of entertainment video title where a user selects a particular video that they desire, for example, to have streamed over IP networks or other networks to home PCs or other display devices. Due to the availability of large numbers of titles available from the network, users of the service require support for narrowing down the space of potential selections. A particular subset of category selection may be performed resulting, for example, in a display of Figure 20 , 400, in which video key frames from designated movie favorites are presented as dynamic collage. At any time in this display, the user may select, as by a mouse click, a key frame of interest which results in the transfer of control to another application such as 306 to a new screen.
Referring now to Figure 23, shown is an example of an embodiment of the screen displayed when the selection of a movie title has been made from the screen 400 of Figure 20 using the browsing tool 316. The display 402 of Figure 23 presents a movie trailer and other information about an associated title selected. Thus, using the browsing tool techniques and display format previously described in combination, the user is assisted in "drilling down quickly" through a large space of movie titles to a point where a choice of whether to select or pay for a movie delivery can be reached. It should be noted that techniques by which representative video key frames may be generated from video content either automatically or semi- automatically are well-known to those schooled in the art.
Using the foregoing as may be done in connection with Figures 22 and 23, a sequence of video titles has been specified. In accordance with techniques previously described herein, server software may dynamically generate an HTML file that represents the results via a query resulting in the title skimming screen, for example, as shown in conjunction with Figure 22 400. The browsing tool 316 takes as input the HTML or XML file that specifies the sequence of title objects, each of which is associated with a sequence of images that may be specified, for example, as URLs previously described in conjunction with the hierarchical data file 150 of
Figure 2A. Also as previously described, as part of the off-line processing, previously stored images, such as JPEG and GIF format files have been made available and accessible using various protocols such as HTTP. As the user views the title sequence, images are loaded incrementally in conjunction with the various display formats chosen. Any one of a variety of different layout techniques may be used, such as those previously described in conjunction with other figures herein.
Referring now to Figure 24, shown is an example of a software presentation player, for example, called DejaView. Generally, as will be described, this DejaView player is a software application that runs as a client on a user's computer in which the user experiences, for example in this application, a virtual auditorium- style presentation that includes temporally synchronized multimedia elements, such as the four media streams previously described, including view graphs, user notes, and audio and video streams of the speaker. The DejaView player such as 304 included in the system 300 of Figure 15, is a tool that allows the user to navigate around a presentation in a random access fashion while maintaining synchronization between the various media streams. The DejaView application may use temporal and spatial browsing techniques of the browser 316 in which view graph miniatures serve as the indices displayed by the browser 316. The view graph images may be generated and stored prior to processing by the browsing tool 316. By selecting a particular view graph miniature displayed in the browser tool 316, control may be transferred to the DejaView player 304 to a particular presentation. Figure 22 is an example of an embodiment of the user interface display of the multimedia presentation using the browsing tool in which miniature view graphs as part of a speaker's presentation are used as indices and displayed. When the user makes a particular selection of a view graph, such as 404a of Figure 22, runtime or execution control may be transferred to the DejaView player 304.
Referring now to Figure 25, shown is an example of the embodiment of a screen displayed when a selection of a multimedia presentation has been made using the browsing tool such as by selecting the miniature view graph 404a. Figure 25 displays an example of a screen under the control of the DejaView player 304 when a particular view graph is selected. The user is presented with the view graph in an enlarged form. Additionally, using the time stamp information associated with the view graph, the user is fast-forwarded to different places corresponding to the time stamp which are synchronized with the time stamp in the associated media streams. In other words, view graph 404a which has been selected occurs at a particular point in the presentation associated with a first time stamp. The first time stamp represents a marker or a "book mark" into the other video streams: the video stream, audio stream and the like. The DejaView player provides, for presenting to the user, a point of synchronization in each of the media streams. Referring now to Figure 26, shown is an example of an embodiment of a user interface display of a multimedia shopping title in which the titles are displayed using a browsing tool 316. In this example embodiment, the proposed technique for indexing into multimedia presentations includes online all e-commerce shopping. Vendors such as QVC have demonstrated consumer appeal for convenience shopping services through the likes of QVC's product showcase television products. In such shows, consumers may be presented of blocks of product descriptions as well as certain related information such as real-time number of sales for that product and the like The internet is one media which may provide similar on-demand access to such services, for example, as allowing displaying of video for product demonstration for a particular product This technique, for example, may enhance traditional television shopping experience through the use of multimedia techniques Various information is displayed in the screen 408 regarding personal computers If a user selects, for example, element 408a from the screen, which is of a particular computer product processor, an e-commerce application may be invoked Referring to Figure 27, shown is an example of an embodiment of a screen display when the selection of a multimedia shopping title 408a has been made using the browsing tool 316 The screen displayed m Figure 27 is that associated with a particular other application such as 306 where control has been transferred from the browser 316 Generally, a DejaView Player 304 is one of the applications that may be used to display, for example, a multimedia presentation selected by a user in conjunction with the browsing tool 316 What will now be described is one embodiment of this player and related components which synchronize various media streams Other techniques may be used in conjunction with other embodiments and implementations Additionally, these techniques used to synchronize the four streams may be used to synchronize a subset of these streams with other applications
The DejaView player 304 is part of a software application that may be used to capture and play multimedia information In particular, this may be used for real time capture and creating synchronized multimedia streams therefrom which may accessed using a communications connection, such as one of a variety of network connections
The capture of the multiple multimedia streams may generally refer to media journaling techniques for capturing events, such as a lecture style presentation with visual aids m real time as they occur and then processing the combination of streams to create a synchronized multimedia presentation. The DejaView player 304 may be described as part of a media journaling application.
Referring now to Figure 28, shown is an example of an embodiment of a media journaling application that includes the player 304 and other application components. This journaling application, in functional terms, provides a user with the capability of capturing multimedia content and related information to create a synchronized multimedia presentation. For example, a seminar style presentation may be captured digitally in real-time. Using certain post-processing techniques, within a very short period thereafter, additional information is combined and synchronized therewith. The new presentation is then made available on a network such as the Internet through a media journal player. The player 304 is, for example, a web browser-based and allows the user to view the whole presentation, or allows for seeking to different places in the presentation while maintaining the synchronization. Various aspects and components will be described in following paragraphs.
Referring to Figure 28, shown is an embodiment of a media journaling system 1100 which includes an audio/video capture digitizer 11 10 that works within a master capture 1120 to create and capture an audio/video presentation for further combination with other inputs, for example, a viewer's notes or a presenter's slides. The audio/video master capture 1110 includes an audio/video capture 1112 which is adapted to receive, for example, an analog video or audio signal, or both (hereinafter referred to as an audio/video signal). The captured audio/video signal is converted to corresponding digital signals that are further compressed in the audio compressor 11 14 and video compressor 11 16. The audio/video capture and digitizer 1110 are well known in the art, and are commercially available, for example, as a Winnov
Videum AV card.
In this embodiment, the master capture 1120 includes a camera/microphone 1122 for actually recording an audio/video event. The output of the camera/microphone 1 122 is typically an analog output that would require digitization and compression by the audio/video capture and digitizer 1110.
However, one of ordinary skill in the art recognizes that with the increased availability of digital audio/video recording devices, audio/video interleaved (AVI) output may be immediately available such that the audio/video capture and digitizer may be bypassed in its entirety.
Time stamps may be determined, for example, in an automated or manual fashion for each of the viewgraphs in which each time stamp is a temporal marker in the speaker presentation. In one embodiment, this time stamp recordal may be automated using a user interface as may be included, for example, in the journaling tool 1 124. Using this tool, the user may select a particular viewgraph to be used as an index. This selection may be made, for example, by clicking a mouse button. Upon this selection, corresponding time stamp information may be recorded in a predetermined file format. One or more of these timestamps may be stored in a timestamp file. The timestamp file may be used in connection with an AVI file, for example, which interleaves audio and video information into a single file format. An AVI file may be produced, for example, by monitoring real-time capture of raw audio-video feed. In one embodiment, an NTSC analog video feed from a camera may be connected to a personal computer. The personal computer may be, for example, a Pentium-class processor having a video digitizer card to digitize the incoming audio/video stream in real time to produce a single interleaved stream, an AVI file. Additionally, other information, such as textual notes, may also be generated using similar manual and/or automated techniques.
In one embodiment, the journaling tool 1 124 used to record the presentation information may be, for example, incorporated into a standard web browser such as Microsoft Internet Explorer or Netscape Navigator. As known to those skilled in the art, alternatives, such as a stand-alone solution, may also be used in which standalone applications may access the Internet or other network directly. The journaling tool may use, for example, Windows Media Technologies from Microsoft Corporation, to integrate timestamp information that may be recorded also using the journaling tool as previously described.
Each of the different media streams, such as the AVI file, may be indexed with the time-stamp information, as may be gathered and stored in the time-stamp file. In one embodiment, a commercially available Microsoft tool, vidtoasf, may be used to produce a file that includes the time stamp information integrated into the AVI file. This tool outputs such a file in a commercially specified file format, the ASF or Active Stream File format that may be served or provided by a software server, for example, written using Microsoft Windows Media Technologies. The ASF file has the time stamp information embedded in it such that a DejaView player may trap and respond to viewgraph change events as the ASF stream the player is receiving plays out. The player then interprets these events using scripting logic within DejaView that updates the slideshow, notes and controls of the tool. In other words, the ASF file may be received by the DejaView player. As the ASF stream is received, the Windows Media Player ActiveX controls used by the player to present the stream is able to raise events corresponding to the timestamps detected in the stream. These events may be trapped and interpreted by scripting logic included in the player.
As known to those skilled in the art, ActiveX controls are generally reusable pieces of code written for example, in C++, Java, Visual Basic and the like. These may be incorporated into an HTML page to extend browser capabilities. Microsoft makes available a variety of ActiveX controls, such as those installed with Internet Explorer. Controls used in one embodiment of DejaView, for example, include Microsoft's Windows Media Technologies Control and Image Control. The Windows Media Player ActiveX control functions as a player of streaming audio/video, embedded in a Web page. It should be noted that viewgraph images may be displayed using the Image ActiveX control using, for example, the PicturePath property which is the URL of a .GIF or .MPEG image to be displayed. An embodiment of the journaling tool may either be physically co-located or physically remote to the journaled event location. If co-located, the journaling tool may include a digitizing PC, video source (from camera), and audio input (local microphone). Available video capture software located on the digitizing PC, includes VidCap from Microsoft Corporation (which does not compress audio) or Premiere from Adobe. When the journaling tool is physically remote from the journaled event, the journal event's audio/video feed may be supplied in analog form via a cable, or digitally via a data network, for example. Using Microsoft's Windows Media Technologies, an AVI file, and the timestamp log file, the ASF file may be produced including the integrated indexing information The ASF file is in a form that may be streamed from a Windows Media Technologies server in real time to the journaler player, the DejaView player 304 The ASF movie may be stored in the Windows Media Technologies server 1 140 along with other media on a web server and is ready for streaming from the server 1 140 It should be noted that the Windows Media Technologies server is currently available from Microsoft Corporation The ASF movie includes the timing information, as may be recorded using a journaling tool A journaling playback device, such as the player 304, may be able to trap and respond to viewgraph change events as they occur accordingly updating other data streams, such as notes The slide miniatures may be located, for example, also on the server 140 that may optionally be connected to a journaling tool
Viewgraph miniatures may also be included m this embodiment, for example, to be displayed by the browser tool 316 These miniatures may be created using Microsoft Powerpoιnt97 if the viewgraphs are created using this application also This may be done, for example, by using the the Powerpoιnt97 "Save as HTML" wizard to generate the JPEG images
In one embodiment as depicted in the system 1 1 10, the journaling software application may include a player and a server component The DejaView player 304 may be the client portion of the application residing and executing on a user's computer The DejaView Player may be written, for example, using Windows Media Technologies and other components as described elsewhere herein The DejaView Server portion may be written using the commercially available Microsoft Windows Media Technologies Server, for example, in which the server computer may be running Windows NT Server 4 0 To support the media capture process, a Pentium class PC running Windows 95 or Windows NT 4 0, for example, may be used A video digitizer card may be, for example, the Winnov Videum AV card that also supports audio capture A video camera, such as a SONY Hι-8 CamCorder, may be used to capture the audio/video feed of a presenter or speaker In this embodiment, other cameras capable of producing an NTSC video out may also be used. It should be noted that the NTSC out of the camera may be fed directly into the video input of the video capture card of the computer supporting realtime video digitization.
The DejaView Player 304 may run, for example, on a Microsoft Windows 95 or NT platform and operates in one embodiment using Microsoft Internet Explorer
4.X. Playback of a DejaView presentation, such as using the ASF file in an embodiment, may be used in a computer system in conjunction with the Microsoft Windows Media Technologies Player. The Windows Media Technologies Player ActiveX control is an integral component in one implementation of the DejaView player. The Microsoft Windows Media Technologies Server may be used to stream audio/video content to DejaView players. The Microsoft Windows Technologies Server is supported, for example, on an NT 4.0 server platform.
Any one of a variety of video capture software tools may be used to digitize DejaView presentation feeds. For example, Microsoft VidCap or Adobe Premiere may be used. Alternatively, different hardware, such as the Videum video capture hardware, may also include software utilities for this purpose. Generally, the foregoing may vary in accordance with each embodiment and implementation. The ASF file may be made available to clients from a Windows Media Technologies server. Other relates files, such as the JPEG viewgraphs, may be placed in locations on a Webserver being used in support of the DejaView player
304.
Referring now to Figure 29, shown is an example of an embodiment of user interface screen of a journal player 304. In this example, the player 400 is an example of an embodiment of the DejaView player 304. It is depicted implemented as a web-based application, and hence is shown as accessed from a predetermined web site 1401. As may be seen, a web command window 1402 is the web based application interface, i.e., a Netscape Navigator or Microsoft Internet Explorer interface for allowing a user to issue commands via the web interface. The journal player 304 allows a user to playback and interact with multimedia presentations of joumaled events. As the ASF movie is received by the journal player 304, Windows
Media Player ActiveX is able to raise events corresponding to the time stamps detected in the movie stream These events may be trapped and interpreted by programming logic of the journal player 304 Importantly, the joumaled event may be provided to the journal player 304 across a network in real time, and a base of available content may change dynamically A talking head window 1403 provides a video window that presents a talking head video of a presenter (with accompanying audio), for example A graphics display window 1404 displays graphic images, for example, the presenter's view graphs An electronic notepad window 1421 displays, for example, textual information pertaining to a presentation A user may also create additional notes while viewing the presentation when a timing event coordinator tool is provided (to be described) Such newly created notes may be kept m synchronization by the journaling tool The notes are storable on network servers for subsequent retrieval, or for sharing amongst users (several sets of notes) in a groupware environment The journal player 304 may present the contents of the talking head window 1403, the graphics display window 1404 and the electronic notepad 1421 at the appropriate times
The journal player 304 may also include a plurality of user controls for allowing a user to navigate the player to different parts of a presentation In the embodiment, the first set of user controls are grouped together as navigation controls 1410 Navigation controls 1410 include a play control 1412, a pause control 1413, and previous slide 1414 and next slide 1415 The play control 1412, as the name implies, causes the presentation to begin to play, either from an initial start of resuming from a pause position The pause control causes the presentation to temporarily freeze until the pause control 1413 is again selected or the play control 412 is elected Alternatively, the play and pause controls 1412 and 1413 are mtegratable into a single control that toggles from the play to pause function and vice versa according to the current function A table of contents menu 141 1 is also included that provides a menu displaying the titles of the view graphs used by the presenter, which menu effectively acts as a table of contents for the presentation A user may thus browse a presentation and go directly to an area of particular interest by selecting the view graph corresponding thereto The presentation will resume synchronized playback from the point of the selected view graph Alternatively, one could add functionality to allow a user to fast forward or fast reverse through the video itself while maintaining synchronization with the associated view graphs and notes.
A second set of user controls are the browse controls 1420 that allow a user to browse the user notes that are displayed in notes window 1421. There may be a one-to-one mapping of the user notes and view graphs in the journaled presentation as ensured by the journaling tool. Using the note window controls 1422 through 1426, the user may move back and forward any number of note pages independently of the rest of the presentation. Forward note control 1422 and back note control 1423 allow the user to go forward and backward through the notes, respectively.
Once a desired note is found, the user may cause the journal player to resynchronize the rest of the presentation with that note by selecting a seek control 1424. A new note control 1425 is provided which allows the user to enter a new note, that may be synched to the presentation at its current position upon selecting the save control 426.
Referring now to Figure 30, shown is a flowchart of steps of a method for capturing and viewing information using the player 304, an embodiment of which has been described as included in Figure 28, and Figure 15. An AVI movie may be created in step 510. This step may additionally be an audio only capture, a video only capture, or an audio/video capture. Alternatively, the AVI movie may take the form of selecting and arranging content from pre-existing sources, such as a video tape of an event. At step 520, time stamps may be captured for any information related to the AVI movie of step 510. As already described, view graphs may be time stamped according to their presentation with a speaker (AVI movie). No limitation is intended in the present invention to view graphs as the AVI movie will dictate what additional content would be co-presented and hence time stamped accordingly. For example, photographs may be presented and time stamped with each introduction during a presentation. As an example, if the joumaled event is a trial, then the exhibits may be time stamped when introduced and discussed. Importantly, the related information need not actually be present at this time because so long as a cue is provided that a time stamp is to be made, the content may be provided later The time stamped related information could come from a web site or server and may be called as necessary during a presentation playback so long as the time stamp and identifying information may be recorded
Having the appropriate video and corresponding related content now being properly time-stamped, a video stream is generated at step 530 with the time stamps now encoded in the AVI movie The images or view graphs are elected in step 540 so that during a playback these images may be located and co-displayed with the presentation In the system 300, these selected viewgraphs may be the multimedia indices, for example, stored in the database 308 and displayed by the browsing tool 316 At step 550 the images and any other items for playback of a presentation are stored on appropriate servers The playback of the presentation on a journal player is accomplished at step 560 Referring back to Figure 15, in system 300, the player 304 may be transferred control by the browsing tool 316 upon the user may requesting a presentation, for example, from a list The images, notes and other information of the presentation may be included in the multimedia information storage 302 This information used in connection with the presentationsmay be accessible from a web site and the information accordingly provided to the journal player with the images, notes, and other non-presentation information being downloaded to the player from their corresponding server locations Microsoft's Windows Media Technologies, as described in other locations herein in one embodiment, is effective for synchronizing the several media elements or software components based on events as they occur, for example, during a video stream However, one limitation is that all time stamp information to be embedded in such video stream must be known a priori, that is, at compile time when the ASF movie is generated While this provides a preferred realization for those instances where interactive synchronization is not required and the convenience of using currently available off-the-shelf products is desired, this does not provide a flexibility to allow users to generate their own notes any time during playback (l e , runtime synchronization) An alternative involves using a timing event coordinator for creating time stamps post video stream Using the foregoing, additional flexibility may be achieved. Additional functionality of the real-time multimedia journaling may be achieved as well by using a timing event coordinator in place of the Windows Media Technologies element. The timing event coordinator keeps time and coordinates events among software componentsthat exhibit temporal state changes. Use of the timing event coordinator herein will be described in an environment for synchronizing components within a multimedia presentation. However, one skilled in the art will recognize that application of the timing event coordinator is not limited to presentation type events. A presentation typically engages a user, has some initial state, T egjn, and changes state at discrete times as the presentation progresses to an end time, Tend- The coordinated components are the separate pieces of the presentation (events) that used to change state at discrete times.
Referring now to Figure 31 , shown is an example of an embodiment of a timeline that may be used to represent time coordination of events in a multimedia presentation. The timeline 601 is an example used in a presentation having only two coordinated components, A and B. However, the following may be generalized to more than two components and associated events. The timeline 601 includes numbered state changes for each component. For example, component A may represent view graph change events having an initial state Ao at the start of the presentation and first changes state at A] . Component B may represent note change events, for example, and may have an initial state B0 and first changes state at Bj . A synchronized application may be created by having each component maintain its own time relative to the start of the presentation, TREL- Alternately, each component may rely on a dedicated timing coordinator for notification to change states. This approach is described in more detail below.
The timing event coordinator may maintain state information for each component for the set of times, S, each relative to the presentation start time, TREL, as to when they should change their state. Thus for component A, S = {A0, Ai, A2, A3, A4} . In addition to registering a set of time events with each component, a type may be used to differentiate the components, i.e., A = view graphs, B = notes. Each type has an index associated with each time, which index identifies the sequential relationship or order (the first state for that type, the second state, etc.). Additionally, each type may have a flag associated therewith to provide information as how to treat that type. As an example, the flag may be binary state of a "0" or a "1" that differentiates between two different treatments in which "0" means treat special, and "1" means no special treatment. Special treatment, in this example, may be that if a user jumps to a certain location in a presentation, then that component type with the "special flag" set may be treated so as the most current component relative to that location, should be displayed.
Thus, when each component registers its set, S, the timing event coordinator may merge S into a table of triples having value, type and index for each element of
S, for which a flag may also be associated. For the timeline provided in FIG. 6, the table of combined sets, C, is as follows:
C = { (0,1,0) (0,2,0) (5,1,1) (21 ,1,2) (30,2,1) (35, 1,3) (46,1,4) (46,2,2) (50,2,3)}.
The first two triples of C describe components A and B respectively and refer to their initial values at time zero, with A represented as a type one, and B represented as a type two, each having an index initially at zero. The third triple in C is associated with component A for a time equal to five seconds, and has an index of "1" (the second A in the set).
The timing event coordinator supports an interface that allows a component to set the time TREL to an arbitrary time value, V. If this occurs, the timing event coordinator looks in C for a matching time value. If a match is found, events are generated with the corresponding type and index fields. These events are used to instruct the component represented by the associated type to perform that action at that time. If no matches are found, the timing event coordinator looks to the next greater triple in C. calculates the difference between the current TREL and sets a timer to raise an event at this time difference. Each time and event is raised by the timer, the period to the next event is determined for the next greater time value and the timer is set accordingly. This is repeated for each next greater time value. A reference for the timer concerning where in the video stream the event occurs is maintained by creating a series of predetermined time stamps in the video stream. In other words, a time stamp may be created every second with no component or event associated therewith. The granularity of such predetermined time stamps is determined according to the specific application requirements. In one embodiment, the timing event coordinator may be implemented as an
ActiveX control and may be used within Internet based applications including software components based on Java applets or ActiveX controls. Neither Java applets or ActiveX controls are intended to be a limitation in an embodiment or implementation of the timing event coordinator. In a typical application of the timing event coordinator, VBScript code within the application registers view graph change events with the timing event coordinator. The registration is with the type = "1," flags and the set, S] = {to, tj, t2, ... tn} of time stamp values. These time stamp values represent the times relative to the start of the presentation (as defined by the audio/video record of the presentation) when view graph change events occurred. For example, t would represent the time when the fifth view graph change occurred.
The notepad component may register with the timing event coordinator with type = 2, flags and the set, S2 = {u0, uj, u2, ... un} of time stamp values. These time stamp values represent the times relative to the start of the presentation that the individual notes were submitted. The timing event coordinator merges the notepad component registered timestamp information with the previously registered view graph display component time stamp information.
For example, if a user selects "goto note three" the application notifies the timing event coordinator and instructs it to go to type = two, index = two. The timing event coordinator responds by generating two events, one to the notepad component requesting a change to note three, and one to the view graph display component requesting a change to the view graph that present at this point in the presentation. Additionally, the application invokes an operation on the timing event coordinator to get the current TREL at the note three point. This is used to inform the audio/video component where to seek to in the associated media stream. Unless otherwise directed by the user, the application proceeds in normal fashion through the remainder of the presentation. Using the timing event coordinator, users may be able to make new notes at runtime (in addition to those already recorded). A user viewing a presentation may take a note at any given point in the presentation. The notepad component may then contact the timing event coordinator to retrieve the current TREL at the new note point. After the note is complete, the notepad component may notify or signal the timing event coordinator and re-register its timing information using the same type as before. This new note then may function as the other notes. Still further, other users may make their set of notes available, for example, from a server, such that a group of users may share notes. Referring now to Figure 32, shown is a flowchart of method steps of an embodiment for synchronizing interrelated components of a multimedia presentation. In this flowchart, steps that may be performed by a timing event coordinator are shown in flow diagram form for synchronizing interrelated components in a presentation as described above. The timing event coordinator method begins at step 701 which represents either the start of a presentation, or is caused by a user request to go to a certain time or event. At step 703 a test is made to determine whether any of the registered software component types have a special flag associated with them. Thus, if any one of the software components has a special flag, the result of the test will always be yes and control will flow to step 705. Assuming that a special flag is associated with one of the software components, a second test as to whether any special processing is needed is made at step 705. An example of special processing, for example, includes determining whether one or more software events need to be updated (or not updated) to the requested command and its associated time. A certain application of a joumaled presentation may require that a view graph be updated to the requested time or command, but not the associated note, hence requiring special processing. Step 707 carries out such special processing.
If no special flags were detected in step 703, processing goes directly to step 709 where a test is made for matches in the relative time, TREL, and the software components. A lack of any software component matches causes processing to proceed to step 71 1 wherein the presentation is restarted at TREL and a check is made to determine whether there is a greater time value, V, left in the presentation. A greater time value V will be treated in step 713 such that the difference between V and TREL is calculated and the timer is set to expire in that calculated time period. Control then returns to step 701. If there is not a greater time value, V, then the playback is concluded. Going back to step 709, if a match was found for TREL and the software component, then step 719 is executed where an event is generated such that the index for the type and index for that matching value. Step 719 is performed if more than one match was found (i.e., a note and view graph changed simultaneously). Steps 717 and 719 are repeated until all matches are processed before control is given to step 711.
While the invention has been particularly shown and described with reference to preferred embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention. The invention has been described using several commercial off the shelf products, which are not meant to be limitations as alternative products may be substituted given they provide suitable functionality for performing the tasks. Still further, the media journaler has been described using a presentation as the application environment. The use of the media journaler may be generalized and apply to other presentations and applications having different event types and occurrences than, for example, coordinating the events of presenters, view graphs and notes.
It should be noted that other techniques may be used to gather time-stamp and other information needed to appropriately synchronize the various media streams in accordance with each embodiment and implementation.

Claims

Claims:
1. A method executed in a computer system for selecting a multimedia presentation comprising: determining a subset of multimedia presentations in accordance with predetermined criteria; determining one or more multimedia data items, each of said one or more multimedia data items being an index into a corresponding one of said subset of multimedia presentations; viewing said one or more multimedia data items using a browser; controlling direction and speed of said one or more multimedia data items viewed; selecting a first of said one or more multimedia data items; and transferring control to machine executable code associated with a first of said subset of multimedia presentations having said first multimedia data item as an index.
2. The method of Claim 1, further comprising: presenting said first multimedia presentation by executing said machine executable code.
3. The method of Claim 1 , wherein said first multimedia presentation is a for on-line shopping of a product.
4. The method of Claim 1 , wherein said first multimedia presentation is a video for video-on-demand selection.
5. The method of Claim 1 , wherein said first multimedia presentation is a speaker presentation using a plurality of media streams and a first of said plurality of media streams includes said first multmedia data item.
6. The method of Claim 5, wherein said first multimedia data item is a miniature viewgraph indexing into other media streams that include an audio file, an image file of speaker notes, and a video file of a speaker giving a presentation.
7. The method of Claim 1, wherein said viewing one or more multimedia data items includes: presenting a hierarchical description of the information in an outline area; and presenting said one or more multimedia data items in a presentation area, said hierarchical description being synchronized with a first temporal arrangement of said one or more multimedia data items in said presentation area.
8. The method of Claim 7, wherein said multmedia data items are presented in said presentation area using a stacking technique in which successive multimedia data items are presented slightly overlapping one or more previously presented multimedia data items, and the method further includes: controlling speed and direction of said one or more multmedia data items presented by stacking and unstacking successively presented multimedia data items in said presentation area.
9. The method of Claim 7, wherein said multimedia data items are presented in said presentation area using a scrolling technique in which successive images are presented along a two-dimensional axis, and the method further includes: controlling speed and direction of said one or more multimedia data items are presented by controlling the direction and speed at which said one or more multimedia data items are presented along said two-dimensional axis.
10. The method of Claim 7, wherein said multimedia data items are presented in said presentation area using a three-dimensional presentation technique in which said multimedia data items are presented with a three-dimensional perspective with regard to a display viewpoint at a particular time.
1 1. The method of Claim 7, wherein some of said multimedia data items includes a group of one or more images having a parent-child relationship in which there is one parent multimedia data item and one or more child multimedia data items, and the method further includes: dividing said presentation area into four quadrants; presenting said parent multimedia data item near a center of said presentation area; and presenting subsequent child multimedia data items beginning in an upper left quadrant and proceeding to present successive multimedia data items on a clockwise rotation in successive quadrants.
12. The method of Claim 1 1 , wherein each of said four quadrants is further divided into four subquadrants, each of said four quadrants having an outermost subquadrant corresponding to an outermost comer of said quadrant with respect to said presentation area, each of said child multimedia data items being presented in an outermost subquadrant associated with a quadrant, an outer comer of said each child multimedia data item being randomly located within said outermost subquadrant.
13. The method of Claim 1, wherein each of said multimedia presentations includes at least two media streams, a first media stream being used and index into said second media stream, wherein said first and second media streams are different.
14. The method of Claim 1 , wherein said second media stream is an audio stream and said first media stream is an image-based medium.
15. The method of Claim 1, wherein each of said first and second media streams are the same.
16. The method of Claim 1 further comprising: producing a database of indices, each of said indices being a multimedia data item.
17. A method executed in a computer system for selecting a multimedia presentation comprising: determining one or more indices for a multimedia presentation, said multimedia presentation having a first media stream and a second media stream, a portion of said first media stream being a first index into a portion of said second media stream, said first index being one of said one or more indices; storing information about said one or more indices in a database; displaying multimedia objects in accordance with said information about said indices in said database using a browser; controlling direction and speed of said multimedia objects displayed; selecting a first of said multimedia objects; and invoking a software program for presenting said multimedia presentation having said first multimedia object as an index.
18. The method of Claim 17 further comprising: wherein said software program is a video on demand application, said first multimedia object is a video key frame, and said multimedia presentation includes a video media stream and an audio media stream.
19. The method of Claim 18, wherein said first media stream is said video media stream and said second media stream is said audio media stream.
20. The method of Claim 17, wherein said multimedia presentation includes a viewgraph stream, a video stream, an audio stream, and a text stream.
21. The method of Claim 20, wherein said viewgraph stream is said first media stream having a portion used as an index into said video stream, said audio stream and said text stream. 22 The method of Claim 21 , wheiem said text stream includes speaker notes, said video stream includes images of a speaker, said audio stream includes voice recordings of said speaker, and said viewgraph stream includes speaker slides 23 The method of Claim 22, wherein a portion of said viewgraphs are used to uniquely differentiate portions of a presentation by said speaker, said viewgraphs being used to synchronize and index into the other media streams.
24. The method of Claim 17, wherein said software program is an electronic commerce shopping application, said first media stream and said second media stream reference a single media stream, said single media stream being an image stream of products for sale.
25. A method executed in a computer system for selecting a multimedia presentation comprising: determining a subset of multimedia presentations in accordance with predetermined criteria, determining one or more multimedia data items, each of said one or more multimedia data items being an index into a corresponding one of said subset of multimedia presentations, viewing said one or more multimedia data items using a browser to select a multimedia presentation, controlling direction and speed of said one or more multimedia data items viewed, and selecting a first of said one or more multimedia data items associated with said multimedia presentation
26 The method of Claim 25, further including transferring control to machine executable code associated with a first of said subset of multimedia presentations having said first multimedia data item as an index
27. A computer program product for selecting a multimedia presentation comprising: machine executable code for determining a subset of multimedia presentations in accordance with predetermined criteria; machine executable code for determining one or more multimedia data items, each of said one or more multimedia data items being an index into a corresponding one of said subset of multimedia presentations; machine executable code for viewing said one or more multimedia data items using a browser; machine executable code for controlling speed and direction of said one or more multimedia data items viewed; machine executable code for selecting a first of said one or more multimedia data items; and machine executable code for transferring control to machine executable code associated with a first of said subset of multimedia presentations having said first multimedia data item as an index.
28. The computer program product of Claim 27, further comprising: machine executable code for presenting said first multimedia presentation by executing said machine executable code.
29. The computer program product of Claim 27,wherein said first multimedia presentation is a for on-line shopping of a product.
30. The computer program product of Claim 27, wherein said first multimedia presentation is a video for video-on-demand selection.
31. The computer program product of Claim 27, wherein said first multimedia presentation is a speaker presentation using a plurality of media streams and a first of said plurality of media streams includes said first multmedia data item.
32. The computer program product of Claim 31, wherein said first multimedia data item is a miniature viewgraph indexing into other media streams that include an audio file, an image file of speaker notes, and a video file of a speaker giving a presentation.
33. The computer program product of Claim 27, wherein said machine executable code for viewing one or more multimedia data items includes: machine executable code for presenting a hierarchical description of the information in an outline area; and machine executable code for presenting said one or more multimedia data items in a presentation area, said hierarchical description being synchronized with a first temporal arrangement of said one or more multimedia data items in said presentation area.
34. The computer program product of Claim 33, wherein said multmedia data items are presented in said presentation area using a stacking technique in which successive multimedia data items are presented slightly overlapping one or more previously presented multimedia data items, and the computer program product further includes: machine executable code for controlling speed and direction of said one or more multmedia data items presented by stacking and unstacking successively presented multimedia data items in said presentation area.
35. The computer program product of Claim 33, wherein said multimedia data items are presented in said presentation area using a scrolling technique in which successive images are presented along a two-dimensional axis, and the computer program product further includes: machine executable code for controlling speed and direction of said one or more multimedia data items are presented by controlling the direction and speed at which said one or more multimedia data items are presented along said two- dimensional axis.
36. The computer program product of Claim 33, wherein said multimedia data items are presented in said presentation area using a three-dimensional presentation technique in which said multimedia data items are presented with a three-dimensional perspective with regard to a display viewpoint at a particular time.
37. The computer program product of Claim 33, wherein some of said multimedia data items includes a group of one or more images having a parent-child relationship in which there is one parent multimedia data item and one or more child multimedia data items, and the computer program product further includes: machine executable code for dividing said presentation area into four quadrants; machine executable code for presenting said parent multimedia data item near a center of said presentation area; and machine executable code for presenting subsequent child multimedia data items beginning in an upper left quadrant and proceeding to present successive multimedia data items on a clockwise rotation in successive quadrants.
38. The computer program product of Claim 37, wherein each of said four quadrants is further divided into four subquadrants, each of said four quadrants having an outermost subquadrant corresponding to an outermost comer of said quadrant with respect to said presentation area, each of said child multimedia data items being presented in an outermost subquadrant associated with a quadrant, an outer comer of said each child multimedia data item being randomly located within said outermost subquadrant.
39. The computer program product of Claim 27. wherein each of said multimedia presentations includes at least two media streams, a first media stream being used and index into said second media stream, wherein said first and second media streams are different.
40. The computer program product of Claim 27, wherein said second media stream is an audio stream and said first media stream is an image-based medium.
41. The computer program product of Claim 27, wherein each of said first and second media streams are the same.
42. The computer program product of Claim 27 further comprising: machine executable code for producing a database of indices, each of said indices being a multimedia data item.
43. A computer program product for selecting a multimedia presentation comprising: machine executable code for determining one or more indices for a multimedia presentation, said multimedia presentation having a first media stream and a second media stream, a portion of said first media stream being a first index into a portion of said second media stream, said first index being one of said one or more indices; machine executable code for storing information about said one or more indices in a database; machine executable code for displaying multimedia objects in accordance with said information about said indices in said database using a browser; machine executable code for controlling speed and direction of said multimedia objects displayed; machine executable code for selecting a first of said multimedia objects; and machine executable code for invoking a software program for presenting said multimedia presentation having said first multimedia object as an index.
44. The computer program product of Claim 43 further comprising: wherein said software program is a video on demand application, said first multimedia object is a video key frame, and said multimedia presentation includes a video media stream and an audio media stream.
PCT/US2000/013561 1999-05-26 2000-05-17 Synchronized spatial-temporal browsing of images for selection of indexed temporal multimedia titles WO2000073914A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU48553/00A AU4855300A (en) 1999-05-26 2000-05-17 Synchronized spatial-temporal browsing of images for selection of indexed temporal multimedia titles

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US13600299P 1999-05-26 1999-05-26
US60/136,002 1999-05-26
US13768899P 1999-06-04 1999-06-04
US60/137,688 1999-06-04
US09/560,006 US7281199B1 (en) 1999-04-14 2000-04-27 Methods and systems for selection of multimedia presentations
US09/560,006 2000-04-27

Publications (1)

Publication Number Publication Date
WO2000073914A1 true WO2000073914A1 (en) 2000-12-07

Family

ID=27384804

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2000/013561 WO2000073914A1 (en) 1999-05-26 2000-05-17 Synchronized spatial-temporal browsing of images for selection of indexed temporal multimedia titles

Country Status (2)

Country Link
AU (1) AU4855300A (en)
WO (1) WO2000073914A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1282049A2 (en) * 2001-07-13 2003-02-05 Samsung Electronics Co., Ltd. System and method for providing summary video information of video data
WO2003017662A1 (en) 2001-08-15 2003-02-27 Sony Corporation Content providing device, content providing method, stram contrent reproducing program, and recorded medium on which stream content reproducing program is recorded
US6928655B1 (en) * 1999-12-16 2005-08-09 Microsoft Corporation Live presentation searching
US7216295B2 (en) 2001-12-20 2007-05-08 Canon Kabushiki Kaisha Method of automatic production of image presentations
US8204750B2 (en) 2005-02-14 2012-06-19 Teresis Media Management Multipurpose media players
US8577683B2 (en) 2008-08-15 2013-11-05 Thomas Majchrowski & Associates, Inc. Multipurpose media players

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5822537A (en) * 1994-02-24 1998-10-13 At&T Corp. Multimedia networked system detecting congestion by monitoring buffers' threshold and compensating by reducing video transmittal rate then reducing audio playback rate
US5894333A (en) * 1996-01-30 1999-04-13 Mitsubishi Denki Kabushiki Kaisha Representative image display method, representative image display apparatus, and motion image search appratus employing the representative image display apparatus
US6006241A (en) * 1997-03-14 1999-12-21 Microsoft Corporation Production of a video stream with synchronized annotations over a computer network
US6014137A (en) * 1996-02-27 2000-01-11 Multimedia Adventures Electronic kiosk authoring system
US6044365A (en) * 1993-09-01 2000-03-28 Onkor, Ltd. System for indexing and retrieving graphic and sound data
US6064379A (en) * 1996-06-24 2000-05-16 Sun Microsystems, Inc. System and method for synchronizing presentation of media stream playlists with real time

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6044365A (en) * 1993-09-01 2000-03-28 Onkor, Ltd. System for indexing and retrieving graphic and sound data
US5822537A (en) * 1994-02-24 1998-10-13 At&T Corp. Multimedia networked system detecting congestion by monitoring buffers' threshold and compensating by reducing video transmittal rate then reducing audio playback rate
US5894333A (en) * 1996-01-30 1999-04-13 Mitsubishi Denki Kabushiki Kaisha Representative image display method, representative image display apparatus, and motion image search appratus employing the representative image display apparatus
US6014137A (en) * 1996-02-27 2000-01-11 Multimedia Adventures Electronic kiosk authoring system
US6064379A (en) * 1996-06-24 2000-05-16 Sun Microsystems, Inc. System and method for synchronizing presentation of media stream playlists with real time
US6006241A (en) * 1997-03-14 1999-12-21 Microsoft Corporation Production of a video stream with synchronized annotations over a computer network

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6928655B1 (en) * 1999-12-16 2005-08-09 Microsoft Corporation Live presentation searching
EP1282049A2 (en) * 2001-07-13 2003-02-05 Samsung Electronics Co., Ltd. System and method for providing summary video information of video data
EP1282049A3 (en) * 2001-07-13 2006-03-22 Samsung Electronics Co., Ltd. System and method for providing summary video information of video data
WO2003017662A1 (en) 2001-08-15 2003-02-27 Sony Corporation Content providing device, content providing method, stram contrent reproducing program, and recorded medium on which stream content reproducing program is recorded
EP1420586A1 (en) * 2001-08-15 2004-05-19 Sony Corporation Content providing device, content providing method, stram contrent reproducing program, and recorded medium on which stream content reproducing program is recorded
EP1420586A4 (en) * 2001-08-15 2010-04-21 Sony Corp Content providing device, content providing method, stram contrent reproducing program, and recorded medium on which stream content reproducing program is recorded
US7216295B2 (en) 2001-12-20 2007-05-08 Canon Kabushiki Kaisha Method of automatic production of image presentations
US8204750B2 (en) 2005-02-14 2012-06-19 Teresis Media Management Multipurpose media players
US9864478B2 (en) 2005-02-14 2018-01-09 Thomas Majchrowski & Associates, Inc. Multipurpose media players
US10514815B2 (en) 2005-02-14 2019-12-24 Thomas Majchrowski & Associates, Inc. Multipurpose media players
US11467706B2 (en) 2005-02-14 2022-10-11 Thomas M. Majchrowski & Associates, Inc. Multipurpose media players
US8577683B2 (en) 2008-08-15 2013-11-05 Thomas Majchrowski & Associates, Inc. Multipurpose media players

Also Published As

Publication number Publication date
AU4855300A (en) 2000-12-18

Similar Documents

Publication Publication Date Title
US9235576B2 (en) Methods and systems for selection of multimedia presentations
JP3449671B2 (en) System and method for enabling creation of personal movie presentations and personal movie collections
US6515656B1 (en) Synchronized spatial-temporal browsing of images for assessment of content
US6144375A (en) Multi-perspective viewer for content-based interactivity
US8332886B2 (en) System allowing users to embed comments at specific points in time into media presentation
US6546405B2 (en) Annotating temporally-dimensioned multimedia content
JP5499331B2 (en) Streaming media trick play
JP4240187B2 (en) Image display apparatus and method
US20110060998A1 (en) System and method for managing internet media content
US20140052770A1 (en) System and method for managing media content using a dynamic playlist
US20040201609A1 (en) Systems and methods of authoring a multimedia file
WO2003079220A1 (en) Method and system for creation, delivery, and presentation of time-synchronized multimedia presentations
JPH0937223A (en) System and method for displaying movie in linkage with source information on which the movie is based
JP2007535018A (en) Multimedia digital content visualization and control technology
JP2007036830A (en) Moving picture management system, moving picture managing method, client, and program
JP2004112372A (en) Image processing system
JP6078476B2 (en) How to customize the display of descriptive information about media assets
WO2000073914A1 (en) Synchronized spatial-temporal browsing of images for selection of indexed temporal multimedia titles
JP4686990B2 (en) Content processing system, content processing method, and computer program
KR20210124906A (en) System And Method For Obtaining Image Information, And Method For Displaying Image Acquisition Information
JP4269980B2 (en) Content processing system, content processing method, and computer program
JP4096670B2 (en) Image playback system
JP4006194B2 (en) Information providing apparatus, information providing program, and recording medium recording the program
JP2002007478A (en) Audiovisual contents providing device and method
JP2009500877A (en) Method and system for device independent media transactions

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: JP