US6618058B1 - Editing device and editing method - Google Patents

Editing device and editing method Download PDF

Info

Publication number
US6618058B1
US6618058B1 US09/587,438 US58743800A US6618058B1 US 6618058 B1 US6618058 B1 US 6618058B1 US 58743800 A US58743800 A US 58743800A US 6618058 B1 US6618058 B1 US 6618058B1
Authority
US
United States
Prior art keywords
time
displaying
image data
editing
clip
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US09/587,438
Inventor
Shigeya Yasui
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Corp
Original Assignee
Sony Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Corp filed Critical Sony Corp
Assigned to SONY CORPORATION reassignment SONY CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: YASUI, SHIGEYA
Application granted granted Critical
Publication of US6618058B1 publication Critical patent/US6618058B1/en
Adjusted expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/034Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/30Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
    • G11B27/3027Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording used signal is digitally coded
    • G11B27/3036Time code signal

Definitions

  • the present invention relates to an editing device and an editing method; and more particularly, is suitably applied to an editing system by which respective images of one imaging object, which are imaged from various angles by using a plurality of video cameras, can be edited.
  • the editing device registers images designated sequentially by an operator from image data stored in the video server, as clips, respectively, prepares sequence of the respective clips registered and an editing list designating an editing content such as joining manner, and carries out so-called non-linear editing by controlling the video server on the basis of the editing list.
  • image data as the imaging result of imaging based on a proper time code by using such as a video camera is recorded by the video server together with given text data accompanied with the image data after a given input processing operation.
  • the imaging object was a series of scenes such as a car race or the like
  • an object of the present invention is to provide an editing device and an editing method for allowing a great improvement of efficiency of editing work.
  • an editing device which comprises first displaying means for displaying respective images based on a plurality of imaging results by converting the images to time lines on the basis of time codes added in synchronicity with respective images on an identical time base respectively, second displaying means for displaying respective images corresponding to time codes designated by the time bases among the respective time lines displayed by the first displaying means, third displaying means for displaying an image of a desired part sequentially designated, on the time bases, from the respective image displayed by second displaying means, as a clip respectively, in which the third displaying means selects only the desired clip from clips corresponding to the respective imaging results configuring the desired parts designated.
  • the editing method has a first step of displaying respective images based on a plurality of imaging results by converting the images to time lines on the basis of time codes added in synchronicity with respective images on an identical time base respectively, a second step of displaying respective images corresponding to time codes designated by the time bases among the respective time lines displayed, a third step of displaying images of a desired part sequentially designated, on the time bases, from the respective displayed images as clips respectively, and a fourth step displaying only a clip selected from clips corresponding to the respective imaging results composing the designated desired parts.
  • temporal correlation between clips extracted from each image of the desired part, which is sequentially designated on the time base, from respective images can be easily understood.
  • FIG. 1 is a block diagram showing an overall configuration of an editing system according to this embodiment
  • FIG. 2 is a block diagram showing the configuration of video cameras
  • FIG. 3 is a block diagram showing the configuration of a processor
  • FIG. 4 is an outlined view showing the configuration of a GUI screen
  • FIG. 5 is an outlined view showing the appearance of preparation of an editing list in the GUI screen
  • FIG. 6 is an outlined view showing the appearance of preparation of the editing list in the GUI screen.
  • FIG. 7 is a flow chart showing steps of preparing process of the editing list.
  • reference numeral 1 denotes an overall editing system according to the embodiment of the present invention as a whole: a plurality of video cameras 2 A to 2 D of the imaging device 2 compresses and encodes respective image data D 1 to D 4 prepared by imaging one imaging object from various angles respectively on the basis of such standard as moving picture experts group phase 2 (MPEG 2) and these are stored as encoded data D 5 to D 8 in the video server 3 comprising an audio video (AV) server having a recording and replaying parts of a disk array configuration.
  • MPEG 2 moving picture experts group phase 2
  • the video server 3 receives text information, such as the titles, IDs, interview data, and interview places of respective image data D 1 to D 4 inputted from a text data-inputting part, not illustrated, by an operator, as text data DT 1 to DT 4 .
  • the video server 3 makes sequentially the text data DT 1 to DT 4 of these image data DI to D 4 to data base to store them.
  • the encoded data D 5 to D 8 of respective image data D 1 to D 4 stored in the video server 3 and corresponding text data DT 1 to DT 4 can be read by the operator as needed by using a processor 5 of the editing device 4 .
  • the processor 5 searches corresponding image data D 1 to D 4 based on the data base of the text information of respective image data D 1 to D 4 stored in the video server 3 to read text data DT 1 to DT 4 of the image data D 1 to D 4 selected from searching.
  • the processor 5 instructs the video server 3 to read the encoded data D 5 to DB of respective corresponding image data D 1 to D 4 on the basis of the result of search, display respective corresponding image data D 1 to D 4 on the basis of the encoded data D 5 to D 8 and the text data DT 1 to DT 4 on a monitor unit 7 respectively by dividing in such as the number of images, and display a part of text information on a liquid crystal display 8 .
  • the operator can select the image data of corresponding image data D 1 to D 4 displayed on the monitor unit 7 and the liquid crystal display 8 and the desired image on the basis of the text information.
  • the processor 5 instructs the video server 3 to read the encoded data D 5 to D 8 and the text data DT 1 to DT 4 of the image data D 1 to D 4 and display an image based on the encoded data D 5 to D 8 on the monitor unit 7 respectively by dividing in the number of the images, and display a part of the text information based on the text data DT 1 to DT 4 on the liquid crystal display 8 .
  • the operator can register sequentially the desired part as clips respectively from images of the image data D 1 to D 4 by operating an operating portion 6 on the basis of the text information displayed on the liquid crystal display 8 to designate IN point and OUT point, watching respective images displayed on the monitor unit 7 .
  • the operator can prepare the editing list designating an editing content, such as a way and sequence of jointing respective clips registered by these steps, by using the processor 5 .
  • the data of the editing list (hereafter, the editing list data) prepared by such way is given from the processor 5 to an editing processing system (not illustrated) to execute actual editing processing on the basis of the editing list data in the editing processing system.
  • the video cameras 2 A to 2 D of an imaging unit 2 are respectively configured as shown such as in FIG. 2, and give images of imaging objects imaged by imaging units 2 A 1 , 2 B 1 , 2 C 1 , and 2 D 1 on the basis of control of an external control unit 10 to corresponding recording signal processing units 2 A 2 , 2 B 2 , 2 C 2 , and 2 D 2 as image signals S 1 A, S 1 B, S 1 C, and S 1 D.
  • the recording signal processing units 2 A 2 to 2 D 2 After having carried out given signal processing for given image signals S 1 A to S 1 D, the recording signal processing units 2 A 2 to 2 D 2 give them to respectively corresponding recording media 2 A 3 , 2 B 3 , 2 C 3 , and 2 D 3 such as a magnetic tape, a digital versatile disc (DVD), a hard disk as S 2 A, A 2 B, S 2 C, and S 2 D.
  • recording media 2 A 3 , 2 B 3 , 2 C 3 , and 2 D 3 such as a magnetic tape, a digital versatile disc (DVD), a hard disk as S 2 A, A 2 B, S 2 C, and S 2 D.
  • respective recording media 2 A 3 to 2 D 3 have received a time code signal S 3 from corresponding time code generating units 2 A 4 , 2 B 4 , 2 C 4 , and 2 D 4 , respectively.
  • the time code generating units 2 A 4 , 2 B 4 , 2 C 4 , and 2 D 4 generate time codes respectively synchronized with respective video cameras 2 A to 2 D on the basis that a reference clock signal CLK is given from an external reference clock generating unit 11 by control of the external controlling unit 10 .
  • respective time code generating units 2 A 4 to 2 D 4 comprise clocks. All these have been adjusted to the identical time by synchronizing on the basis of the reference clock signal CLK given from the reference clock generating unit.
  • recording processing signals S 2 A to S 2 D given from the recording signal processing units 2 A 2 to 2 D 2 and a time code signal S 3 given from the time code generating units 2 A 4 to 2 D 4 are integrally recorded as recording signals S 4 A, S 4 B, S 4 C, and S 4 D.
  • the recording media 2 A 3 to 2 D 3 read respectively recording signals S 4 A to S 4 D as needed on the basis of control of the external controlling unit 10 to give them to replaying signal processing units 2 A 5 , 2 B 5 , 2 C 5 , and 2 D 5 .
  • the replaying signal processing units 2 A 5 to 2 D 5 carry out given signal processing for recording signals S 4 A to S 4 D respectively given to give image data D 1 to D 4 yielded to the video server 3 of the editing system 1 .
  • the processor 5 of the editing device 4 is configured by connecting central processing unit (CPU) 20 , read only memory (ROM) 21 , random access memory (RAM) 22 , a display processing circuit 23 , a signal processing circuit 24 , interface circuit 25 and 26 and small computer system interface (SCSI) interface circuit 27 through a CPU bus 28 , connected to an operating portion 6 and the video server 3 respectively through respective interface circuit 25 and 26 , and connected to a hard disk device 29 through SCSI interface circuit 27 .
  • CPU central processing unit
  • ROM read only memory
  • RAM random access memory
  • SCSI small computer system interface
  • the CPU 20 reads image data in the ROM 21 as needed on the basis of a program stored in the ROM 21 , gives this to the liquid crystal display 8 as an image signal S 10 through the display processing circuit 23 , thereby displaying numerals and characters, which are inputted through a GUI screen 30 and the operating portion 6 , on the liquid crystal display 8 .
  • the CPU 20 replays the encoded data D 5 to D 8 of necessary image data D 1 to D 4 and their text data DT 1 to DT 4 by accessing the video server 3 through the second interface circuit 26 as needed on the basis of a program stored in the ROM 21 , when an executable instruction of a given processing is given through the first interface circuit 25 by operation of the operating portion 6 .
  • the CPU 20 takes in the encoded data D 5 to D 8 and the text data DT 1 to DT 4 of the image data D 1 to D 4 replayed by the video server 3 through the signal processing circuit 24 and the second interface circuit 26 and decodes the encoded data D 5 to D 8 on the basis of such standard as MPEG2 in the signal processing circuit 24 to yield decoded data D 9 to D 12 .
  • the CPU 20 displays images of corresponding image data D 1 to D 4 and text information corresponding thereto on the monitor 7 and the liquid crystal display 8 by giving these decoded data D 9 to D 12 and the text data DT 1 to DT 4 to the monitor 7 and the liquid crystal display 8 as respective image signals S 11 and S 12 through the display processing circuit 23 .
  • the CPU 20 stores these decoded data D 9 to D 12 of the image data D 1 to D 4 yielded by such way and the corresponding text data DT 1 to DT 4 in the hard disk device 29 as needed through SCSI interface circuit 27 , and reads these decoded data D 9 to D 12 of the image material and the corresponding text data DT 1 to DT 4 stored in the hard disk device 29 through SCSI interface circuit 27 as needed.
  • the text information displayed practically on the liquid crystal display 8 is, as shown in FIG. 4, configured by graphical user interface (GUI) screen 30 comprising a data displaying area 30 A, in which an image content such as a total imaging time of the respective image data D 1 to D 4 on the basis of the text data DT 1 to DT 4 of the respective image data D 1 to D 4 selected by the operator is displayed in parallel respectively on the identical time base as time data D TIME 1 to D TIME 4 B composed of time lines based on a time code signal S 3 respectively, and a clip displaying area 30 B, in which the desired part of the respective image data D 1 to D 4 designated by the operator in this data displaying area 30 A is displayed in the AB roll style which displays sequentially and alternately in A roll side and B roll side in parallel by using the given time lines respectively as clips.
  • GUI graphical user interface
  • the data displaying area 30 A is time display based on the clock of the video cameras 2 A to 2 D, and therefore, in this embodiment shows the imaging time from about 18:00 to slightly past 21:00.
  • the operator designates an IN point P IN 1 and an OUT point P OUT 1 watching images of the respective image data D 1 to D 4 correspondingly displayed in a four-part split screen on the monitor 7 to extract images of a desired time zone as clip group C 1 from the time data D TIME 1 to D TIME 4 B of the respective image data D 1 to D 4 displayed in the data displaying area 30 A from about 18:00 to slightly past 21:00.
  • the clip group C 1 extracted by these steps is displayed as clips C 1 A, C 1 B, C 1 C, and C 1 D by using the time line of a length corresponded to the time of the time data D TIME 1 to D TIME 4 B in A roll side of a clip displaying area 30 B.
  • the operator designates an IN point P IN 2 and P IN 3 and an OUT point P OUT 2 and P OUT 3 to extract images of a desired time zone as clip groups C 2 and C 3 from the time data D TIME 1 to D TIME 4 B displayed in the data displaying area 30 A.
  • these clip groups C 2 and C 3 are displayed sequentially and alternately as clips C 2 A to C 2 C and C 3 A, C 3 B, and C 3 D on B roll side and A roll side of a clip displaying area 30 B.
  • the length of respective clips C 1 A to C 3 D of clip groups C 1 to C 3 displayed in the clip displaying area 30 B respectively proportions to the time between the IN point P IN 1 and P IN 3 and the OUT point P OUT 1 and P OUT 3 designated in the data displaying area 30 A, and is set longer for a given time (about 5 seconds) back and forth as joint margins for jointing.
  • the operator selects desired images of the clips C 1 A to C 3 D from the respective clip group C 1 to C 3 for every the respective clip group C 1 to C 3 watching images of the clips C 1 A to C 3 D displayed on the monitor 7 corresponding to the respective clip group C 1 to C 3 jointed by arranging in a given alignment in the AB roll system in the clip displaying area 30 B in order to designate the timing of switching the images by using switching points P SW 1 to P SW 4 .
  • moving, adding, and deleting these switching points P SW 1 to P SW 4 in the clip groups C 1 to C 3 jointed allows easy edition keeping a time flow of the clip groups C 1 to C 3 jointed.
  • the CPU 20 of the processor 5 executes the editing processing of the image data D 1 to D 4 described above according to a processing step RT 1 for preparing the editing list shown in FIG. 7 in the mode for preparing the editing list.
  • the CPU 20 starts the processing step RT 1 for preparing the editing list in a step SP 0 and in the following step SP 1 , displays the necessary time for the image data D 1 to D 4 in the data displaying area 30 A of the GUI screen 30 of the liquid crystal display 8 by synchronizing with the identical time base in parallel on the basis of the corresponding text data DT 1 to DT 4 .
  • the CPU 20 goes on to the next step SP 2 , and when an instruction designating a desired part of the image data D 1 to D 4 in the data displaying area 30 A using the IN point P IN 1 and the OUT point P OUT 1 is inputted, extracts this as the clip group C 1 and goes on to the following step SP 3 .
  • the CPU 20 displays the clip group C 1 extracted in the step SP 2 in the clip displaying area 30 B as clips C 1 A to C 1 D by using corresponding time lines in the step SP 3 and goes on to the step SP 4 to decide to complete or not extraction of a clip from the image data D 1 to D 4 in the data displaying area 30 A.
  • the CPU 20 returns to the step SP 2 when a negative result is obtained by input of the designation instruction of the IN point P IN 2 and the OUT point P OUT 2 for extraction of clips by the operator in the step SP 4 , and in the following step SP 4 , repeats a loop of the steps SP 2 -SP 3 -SP 4 until a positive result is obtained, finally goes on to the next step SP 5 when the positive result is obtained in the step SP 4 by input of an instruction to complete extraction of the clip by the operator.
  • the CPU 20 displays sequentially and alternately the clip group C 1 to C 3 extracted by repeating the loop of the steps SP 2 -SP 3 -SP 4 in the clip displaying area 30 B from the A roll side in the AB roll system in order to joint the clip group C 1 to C 3 extracted in a given condition.
  • the CPU 20 puts the switching points P SW 1 to P SW 4 in the point designated in the respective clip group C 1 to C 3 to categorize for each area as well as to point to corresponding clips C 1 A to C 1 D corresponding to images designated in the area.
  • the CPU 20 goes on to the following step SP 6 to complete the processing step RT 1 for preparing the editing list.
  • one imaging object is imaged by adding the time code signal S 3 respectively synchronized from various angles by using a plurality of video cameras 2 A to 2 D.
  • the editing system 1 in the mode for preparing the editing list converts the image data D 1 to D 4 as the result of imaging to the time data D TIME 1 to D TIME 4 B on the basis of the time code signal S 3 respectively in order to display in the data displaying area 30 A of the GUI screen 30 in parallel in the identical time base.
  • the clips C 1 A to C 1 D which are extracted from the time data D TIME 1 to D TIME 4 B displayed in the data displaying area 30 A for every image corresponding to the desired part which is designated sequentially on the time base, are displayed in the clip displaying area 30 B in the AB roll system as the clip group C 1 to C 3 for every clips C 1 A to C 3 D corresponding to the desired part in order to join them.
  • the present invention is not restricted to this, but can be widely applied to other various editing devices in which the temporal correlation between the clips selected from the image data can be easily understood.
  • time code generating unit 2 A 4 to 2 D 4 of the video cameras 2 A to 2 D has been described.
  • the present invention is not restricted to this.
  • Various other types can be widely applied as the time code generating unit 2 A 4 to 2 D 4 , as long as they generate time codes synchronizing respectively based on the reference clock signal CLK.
  • the editing device comprises the first displaying means displaying images based on a plurality of imaging results through converting to time lines with a standard of time codes added synchronizing with respective images on an identical time base respectively, the second displaying means displaying respective images corresponding to time codes designated by the time bases among the respective time lines displayed, the third displaying means displaying an image of a desired part sequentially designated, on the time bases, from the respective image displayed, as a clip respectively, wherein the third displaying means selects only the desired clip from clips corresponding to the respective imaging results configuring the desired parts designated.
  • the temporal correlation between the clips extracted for every image of the desired part, which is designated sequentially from respective images on the time base can be easily understood.
  • the editing system in which the efficiency of editing work can be greatly improved, can be realized.
  • the editing method comprises a first step displaying respective images based on a plurality of imaging results through converting to time lines with a standard of time codes added synchronizing with respective images on an identical time base respectively, a second step displaying respective images corresponding to time codes designated by the time bases among the respective time lines displayed, a third step displaying images of a desired part sequentially designated, on the time bases, from the respective displayed images as clips respectively, and a fourth step displaying only a clip selected from clips corresponding to the respective imaging results composing the designated desired parts.
  • the temporal correlation between the clips extracted for every image of the desired part, which is designated sequentially from respective images on the time base can be easily understood.
  • the editing system in which the efficiency of editing work can be greatly improved, can be realized.

Abstract

An editing device and an editing method which greatly improve the efficiency of editing work. The editing device includes first displaying means for displaying a plurality of images by converting the images to time lines on the basis of time codes added in synchronicity with respective images on an identical time base respectively, second displaying means for displaying images corresponding to the time codes designated on the time bases among the respective time lines, and the third displaying means for displaying an image of a desired part sequentially designated, on the time bases, from the respective images, as a clip respectively. The editing method is configured by displaying a plurality of images by converting the images to time lines on the basis of a time codes added respectively in synchronicity with an identical time base, by displaying the images corresponding to the time codes designated on the time bases among the respective time lines, and by displaying an image of a desired part sequentially designated, on the time bases, from the respective images, as a clip respectively.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates to an editing device and an editing method; and more particularly, is suitably applied to an editing system by which respective images of one imaging object, which are imaged from various angles by using a plurality of video cameras, can be edited.
2. Description of the Related Art
So far, as this kind of editing system, there is one configured by an editing device and a video server; the editing device registers images designated sequentially by an operator from image data stored in the video server, as clips, respectively, prepares sequence of the respective clips registered and an editing list designating an editing content such as joining manner, and carries out so-called non-linear editing by controlling the video server on the basis of the editing list.
Meanwhile, according to the editing system, image data as the imaging result of imaging based on a proper time code by using such as a video camera is recorded by the video server together with given text data accompanied with the image data after a given input processing operation.
Therefore, for example, when one imaging object is imaged from various angles by using a plurality of above-mentioned video cameras, those respective images imaged are imaged based on separate time codes, respectively.
However, there was the following problem in such an editing system that it was difficult to understand temporal correlation among respective images due to desynchronized time sequence of those respective images for editing respective images imaged for one imaging object with respective video cameras by such way.
Therefore, for example, in the case where the imaging object was a series of scenes such as a car race or the like, it was difficult to understand which portion of which scene was the image of a clip selected from respective image data, and this posed a problem that adjusting temporal timing between the clips was difficult.
Further, there is such a problem in preparing an editing list by this editing system that a juncture between clips becomes unclear, because of displaying by jointing selected clips to form a series of time lines on the time base.
SUMMARY OF THE INVENTION
In view of the foregoing, an object of the present invention is to provide an editing device and an editing method for allowing a great improvement of efficiency of editing work.
The foregoing object and other objects of the invention have been achieved by the provision of an editing device which comprises first displaying means for displaying respective images based on a plurality of imaging results by converting the images to time lines on the basis of time codes added in synchronicity with respective images on an identical time base respectively, second displaying means for displaying respective images corresponding to time codes designated by the time bases among the respective time lines displayed by the first displaying means, third displaying means for displaying an image of a desired part sequentially designated, on the time bases, from the respective image displayed by second displaying means, as a clip respectively, in which the third displaying means selects only the desired clip from clips corresponding to the respective imaging results configuring the desired parts designated.
As a result, according to this editing device, temporal correlation between clips extracted from each image of the desired part, which is sequentially designated on the time base, from respective images can be easily understood.
The editing method according to the present invention has a first step of displaying respective images based on a plurality of imaging results by converting the images to time lines on the basis of time codes added in synchronicity with respective images on an identical time base respectively, a second step of displaying respective images corresponding to time codes designated by the time bases among the respective time lines displayed, a third step of displaying images of a desired part sequentially designated, on the time bases, from the respective displayed images as clips respectively, and a fourth step displaying only a clip selected from clips corresponding to the respective imaging results composing the designated desired parts.
As a result, according to the editing method, temporal correlation between clips extracted from each image of the desired part, which is sequentially designated on the time base, from respective images can be easily understood.
The nature, principle and utility of the invention will become more apparent from the following derailed description when read in conjunction with the accompanying drawings in which like parts are designated by like reference numerals or characters.
BRIEF DESCRIPTION OF THE DRAWINGS
In the accompanying drawings:
FIG. 1 is a block diagram showing an overall configuration of an editing system according to this embodiment;
FIG. 2 is a block diagram showing the configuration of video cameras;
FIG. 3 is a block diagram showing the configuration of a processor;
FIG. 4 is an outlined view showing the configuration of a GUI screen;
FIG. 5 is an outlined view showing the appearance of preparation of an editing list in the GUI screen;
FIG. 6 is an outlined view showing the appearance of preparation of the editing list in the GUI screen; and
FIG. 7 is a flow chart showing steps of preparing process of the editing list.
DETAILED DESCRIPTION OF THE EMBODIMENT
Preferred embodiments of this invention will be described with reference to the accompanying drawings:
(1) Configuration of an Editing System According to the Present Embodiment
In FIG. 1, reference numeral 1 denotes an overall editing system according to the embodiment of the present invention as a whole: a plurality of video cameras 2A to 2D of the imaging device 2 compresses and encodes respective image data D1 to D4 prepared by imaging one imaging object from various angles respectively on the basis of such standard as moving picture experts group phase 2 (MPEG 2) and these are stored as encoded data D5 to D8 in the video server 3 comprising an audio video (AV) server having a recording and replaying parts of a disk array configuration.
In this step, the video server 3 receives text information, such as the titles, IDs, interview data, and interview places of respective image data D1 to D4 inputted from a text data-inputting part, not illustrated, by an operator, as text data DT1 to DT4. Through this step, the video server 3 makes sequentially the text data DT1 to DT4 of these image data DI to D4 to data base to store them.
Next, the encoded data D5 to D8 of respective image data D1 to D4 stored in the video server 3 and corresponding text data DT1 to DT4 can be read by the operator as needed by using a processor 5 of the editing device 4.
In practice, in a data searching and selection mode, when search criteria for IDs of the image data D1 to D4 or the like that are desired by the operator are inputted to an operating portion 6 and a searching instruction for the image data D1 to D4 is inputted, the processor 5 searches corresponding image data D1 to D4 based on the data base of the text information of respective image data D1 to D4 stored in the video server 3 to read text data DT1 to DT4 of the image data D1 to D4 selected from searching.
In addition, the processor 5 instructs the video server 3 to read the encoded data D5 to DB of respective corresponding image data D1 to D4 on the basis of the result of search, display respective corresponding image data D1 to D4 on the basis of the encoded data D5 to D8 and the text data DT1 to DT4 on a monitor unit 7 respectively by dividing in such as the number of images, and display a part of text information on a liquid crystal display 8.
As a result, the operator can select the image data of corresponding image data D1 to D4 displayed on the monitor unit 7 and the liquid crystal display 8 and the desired image on the basis of the text information.
When a replaying instruction of the image data D1 to D4 selected as described above is inputted in the mode to prepare the editing list following to data searching and selection mode, the processor 5 instructs the video server 3 to read the encoded data D5 to D8 and the text data DT1 to DT4 of the image data D1 to D4 and display an image based on the encoded data D5 to D8 on the monitor unit 7 respectively by dividing in the number of the images, and display a part of the text information based on the text data DT1 to DT4 on the liquid crystal display 8.
As a result, the operator can register sequentially the desired part as clips respectively from images of the image data D1 to D4 by operating an operating portion 6 on the basis of the text information displayed on the liquid crystal display 8 to designate IN point and OUT point, watching respective images displayed on the monitor unit 7.
Also, the operator can prepare the editing list designating an editing content, such as a way and sequence of jointing respective clips registered by these steps, by using the processor 5. The data of the editing list (hereafter, the editing list data) prepared by such way is given from the processor 5 to an editing processing system (not illustrated) to execute actual editing processing on the basis of the editing list data in the editing processing system.
(2) Configuration of Video Cameras 2A to 2D of an Imaging Unit 2
Practically, the video cameras 2A to 2D of an imaging unit 2 are respectively configured as shown such as in FIG. 2, and give images of imaging objects imaged by imaging units 2A1, 2B1, 2C1, and 2D1 on the basis of control of an external control unit 10 to corresponding recording signal processing units 2A2, 2B2, 2C2, and 2D2 as image signals S1A, S1B, S1C, and S1D.
After having carried out given signal processing for given image signals S1A to S1D, the recording signal processing units 2A2 to 2D2 give them to respectively corresponding recording media 2A3, 2B3, 2C3, and 2D3 such as a magnetic tape, a digital versatile disc (DVD), a hard disk as S2A, A2B, S2C, and S2D.
In this step, respective recording media 2A3 to 2D3 have received a time code signal S3 from corresponding time code generating units 2A4, 2B4, 2C4, and 2D4, respectively. The time code generating units 2A4, 2B4, 2C4, and 2D4 generate time codes respectively synchronized with respective video cameras 2A to 2D on the basis that a reference clock signal CLK is given from an external reference clock generating unit 11 by control of the external controlling unit 10.
For reference, in this example, respective time code generating units 2A4 to 2D4 comprise clocks. All these have been adjusted to the identical time by synchronizing on the basis of the reference clock signal CLK given from the reference clock generating unit.
By this, in the recording media 2A3 to 2D3, recording processing signals S2A to S2D given from the recording signal processing units 2A2 to 2D2 and a time code signal S3 given from the time code generating units 2A4 to 2D4 are integrally recorded as recording signals S4A, S4B, S4C, and S4D.
The recording media 2A3 to 2D3 read respectively recording signals S4A to S4D as needed on the basis of control of the external controlling unit 10 to give them to replaying signal processing units 2A5, 2B5, 2C5, and 2D5.
By this, the replaying signal processing units 2A5 to 2D5 carry out given signal processing for recording signals S4A to S4D respectively given to give image data D1 to D4 yielded to the video server 3 of the editing system 1.
By such steps, in respective video cameras 2A to 2D, images of imaging objects respectively imaged are integrally recorded with time code signal S3 respectively synchronized to be read and replayed as needed.
(3) Configuration of the Processor 5 of the Editing Device 4
In thus step, the processor 5 of the editing device 4, as shown in FIG. 3, is configured by connecting central processing unit (CPU) 20, read only memory (ROM) 21, random access memory (RAM) 22, a display processing circuit 23, a signal processing circuit 24, interface circuit 25 and 26 and small computer system interface (SCSI) interface circuit 27 through a CPU bus 28, connected to an operating portion 6 and the video server 3 respectively through respective interface circuit 25 and 26 , and connected to a hard disk device 29 through SCSI interface circuit 27.
In this case, the CPU 20 reads image data in the ROM 21 as needed on the basis of a program stored in the ROM 21, gives this to the liquid crystal display 8 as an image signal S10 through the display processing circuit 23, thereby displaying numerals and characters, which are inputted through a GUI screen 30 and the operating portion 6, on the liquid crystal display 8.
Further, the CPU 20 replays the encoded data D5 to D8 of necessary image data D1 to D4 and their text data DT1 to DT4 by accessing the video server 3 through the second interface circuit 26 as needed on the basis of a program stored in the ROM 21, when an executable instruction of a given processing is given through the first interface circuit 25 by operation of the operating portion 6.
Furthermore, the CPU 20 takes in the encoded data D5 to D8 and the text data DT1 to DT4 of the image data D1 to D4 replayed by the video server 3 through the signal processing circuit 24 and the second interface circuit 26 and decodes the encoded data D5 to D8 on the basis of such standard as MPEG2 in the signal processing circuit 24 to yield decoded data D9 to D12.
The CPU 20 displays images of corresponding image data D1 to D4 and text information corresponding thereto on the monitor 7 and the liquid crystal display 8 by giving these decoded data D9 to D12 and the text data DT1 to DT4 to the monitor 7 and the liquid crystal display 8 as respective image signals S11 and S12 through the display processing circuit 23.
The CPU 20 stores these decoded data D9 to D12 of the image data D1 to D4 yielded by such way and the corresponding text data DT1 to DT4 in the hard disk device 29 as needed through SCSI interface circuit 27, and reads these decoded data D9 to D12 of the image material and the corresponding text data DT1 to DT4 stored in the hard disk device 29 through SCSI interface circuit 27 as needed.
The text information displayed practically on the liquid crystal display 8 is, as shown in FIG. 4, configured by graphical user interface (GUI) screen 30 comprising a data displaying area 30A, in which an image content such as a total imaging time of the respective image data D1 to D4 on the basis of the text data DT1 to DT4 of the respective image data D1 to D4 selected by the operator is displayed in parallel respectively on the identical time base as time data D TIME 1 to DTIME 4B composed of time lines based on a time code signal S3 respectively, and a clip displaying area 30B, in which the desired part of the respective image data D1 to D4 designated by the operator in this data displaying area 30A is displayed in the AB roll style which displays sequentially and alternately in A roll side and B roll side in parallel by using the given time lines respectively as clips.
For reference, the data displaying area 30A is time display based on the clock of the video cameras 2A to 2D, and therefore, in this embodiment shows the imaging time from about 18:00 to slightly past 21:00.
Subsequently, in the preparation mode for the editing list, as shown in FIG. 5A, the operator designates an IN point P IN 1 and an OUT point P OUT 1 watching images of the respective image data D1 to D4 correspondingly displayed in a four-part split screen on the monitor 7 to extract images of a desired time zone as clip group C1 from the time data D TIME 1 to DTIME 4B of the respective image data D1 to D4 displayed in the data displaying area 30A from about 18:00 to slightly past 21:00.
In this step, the area designated by the IN point P IN 1 and the OUT point P OUT 1 is displayed separately.
The clip group C1 extracted by these steps is displayed as clips C1A, C1B, C1C, and C1D by using the time line of a length corresponded to the time of the time data D TIME 1 to DTIME 4B in A roll side of a clip displaying area 30B.
By repeating these steps sequentially, the operator, as shown in FIG. 5B, designates an IN point P IN 2 and PIN 3 and an OUT point P OUT 2 and POUT 3 to extract images of a desired time zone as clip groups C2 and C3 from the time data D TIME 1 to DTIME 4B displayed in the data displaying area 30A.
By these steps, these clip groups C2 and C3 are displayed sequentially and alternately as clips C2A to C2C and C3A, C3B, and C3D on B roll side and A roll side of a clip displaying area 30B.
In this step, the length of respective clips C1A to C3D of clip groups C1 to C3 displayed in the clip displaying area 30B respectively proportions to the time between the IN point P IN 1 and PIN 3 and the OUT point P OUT 1 and POUT 3 designated in the data displaying area 30A, and is set longer for a given time (about 5 seconds) back and forth as joint margins for jointing.
Subsequently, as shown in FIG. 6, the operator selects desired images of the clips C1A to C3D from the respective clip group C1 to C3 for every the respective clip group C1 to C3 watching images of the clips C1A to C3D displayed on the monitor 7 corresponding to the respective clip group C1 to C3 jointed by arranging in a given alignment in the AB roll system in the clip displaying area 30B in order to designate the timing of switching the images by using switching points P SW 1 to P SW 4.
In this step, the clips C1A to C3D designated by the operator are displayed separately.
For reference, in this editing system 1, moving, adding, and deleting these switching points P SW 1 to P SW 4 in the clip groups C1 to C3 jointed allows easy edition keeping a time flow of the clip groups C1 to C3 jointed.
By these steps in the editing system 1, in extraction of the desired part from the image data D1 to D4 as the clip groups C1 to C3 after the operator selects the desired image data D1 to D4 from image data stored in the video server 3 using the GUI screen 30, a temporal correlation between the clip groups C1 to C3 to be extracted is easily understood because the image data D1 to D4 is arranged in parallel on the identical time base on the basis of the time code signal S3. In addition, the temporal timing between the clip groups C1 to C3 can be easily adjusted, because the clip groups C1 to C3 are synchronized on the basis of the time code signal S3.
(4) Processing Steps for Preparing the Editing List for the CPU 20
Practically, herewith, the CPU 20 of the processor 5 executes the editing processing of the image data D1 to D4 described above according to a processing step RT1 for preparing the editing list shown in FIG. 7 in the mode for preparing the editing list.
When the desired image data D1 to D4 are designated from image data stored in the video server 3 by the operator to input replay instruction thereof, the CPU20 starts the processing step RT1 for preparing the editing list in a step SP0 and in the following step SP1, displays the necessary time for the image data D1 to D4 in the data displaying area 30A of the GUI screen 30 of the liquid crystal display 8 by synchronizing with the identical time base in parallel on the basis of the corresponding text data DT1 to DT4.
Subsequently, the CPU20 goes on to the next step SP2, and when an instruction designating a desired part of the image data D1 to D4 in the data displaying area 30A using the IN point P IN 1 and the OUT point P OUT 1 is inputted, extracts this as the clip group C1 and goes on to the following step SP3.
Next, the CPU20 displays the clip group C1 extracted in the step SP2 in the clip displaying area 30B as clips C1A to C1D by using corresponding time lines in the step SP3 and goes on to the step SP4 to decide to complete or not extraction of a clip from the image data D1 to D4 in the data displaying area 30A.
Next, the CPU20 returns to the step SP2 when a negative result is obtained by input of the designation instruction of the IN point P IN 2 and the OUT point P OUT 2 for extraction of clips by the operator in the step SP4, and in the following step SP4, repeats a loop of the steps SP2-SP3-SP 4 until a positive result is obtained, finally goes on to the next step SP5 when the positive result is obtained in the step SP4 by input of an instruction to complete extraction of the clip by the operator.
In this step, the CPU20 displays sequentially and alternately the clip group C1 to C3 extracted by repeating the loop of the steps SP2-SP3-SP 4 in the clip displaying area 30B from the A roll side in the AB roll system in order to joint the clip group C1 to C3 extracted in a given condition.
Next, when the instruction to select which image in the respective clip group C1 to C3 jointed by the operator in the step SP5 and to designate the timing of switching this is inputted, the CPU20 puts the switching points P SW 1 to P SW 4 in the point designated in the respective clip group C1 to C3 to categorize for each area as well as to point to corresponding clips C1A to C1D corresponding to images designated in the area.
Subsequently, the CPU20 goes on to the following step SP6 to complete the processing step RT1 for preparing the editing list.
(5) Operation and Effect of the Present Embodiment
According to the configuration described above, in the editing system 1, one imaging object is imaged by adding the time code signal S3 respectively synchronized from various angles by using a plurality of video cameras 2A to 2D.
The editing system 1 in the mode for preparing the editing list converts the image data D1 to D4 as the result of imaging to the time data D TIME 1 to DTIME 4B on the basis of the time code signal S3 respectively in order to display in the data displaying area 30A of the GUI screen 30 in parallel in the identical time base.
In the editing system 1, the clips C1A to C1D, which are extracted from the time data D TIME 1 to DTIME 4B displayed in the data displaying area 30A for every image corresponding to the desired part which is designated sequentially on the time base, are displayed in the clip displaying area 30B in the AB roll system as the clip group C1 to C3 for every clips C1A to C3D corresponding to the desired part in order to join them.
Therefore, in the editing system 1, the temporal correlation of the respective clips C1A to C3D is easily understood and also jointing borders between the clip groups C1 to C3 to be jointed can become clear.
According to configuration described above, in the mode of preparing the editing list in the editing system 1, respective images based on a plurality of imaging result is converted respectively to time lines in a reference of the time code added synchronizing with the respective images on the identical time base in order to display in parallel to the data displaying area 30A. By this, the temporal correlation of the respective clips, which are extracted for every image of the desired part, which is sequentially designated on the time base from the respective images, can be easily understood. Thus, the editing system 1, in which the efficiency of editing work can be greatly improved, can be realized.
(6) Other Embodiments
In the embodiments described above, a case of an application of the present invention to the editing system 1 has been described. However, the present invention is not restricted to this, but can be widely applied to other various editing devices in which the temporal correlation between the clips selected from the image data can be easily understood.
Further in embodiments described above, a case where the image data D1 to D4 imaged by using the video cameras 2A to 2D as image data to be stored in the video server 3 has been described. However, the present invention is not restricted to this. Various other types of image data, as long as they are a plurality of image data imaged on the basis of respectively synchronized time code, can be widely applied as the image data to be stored in the video server 3.
Furthermore, in the embodiments described above, a case, where the data displaying area 30A of the GUI screen 30 and the clip displaying area 30B are displayed as shown in FIG. 4, has been described. However, the present invention is not restricted to this, but various other displaying methods can be widely applied as the displaying method of the GUI screen 30.
Furthermore, in the embodiments described above, a case, where divided displaying is carried out for the number of images based on the image data D1 to D4 selected from the image data D1 to D4, which are stored in the video server 3 by the operator, as the displaying method for images in the monitor 7, has been described. However, the present invention is not restricted to this, but various other displaying methods can be applied as the displaying method for images of image data D1 to D4. For example, it is possible to use the monitor 7 for the number of images selected from the image data D1 to D4 to display the images corresponding to each of them.
Further, in the embodiments described above, a case, where a clock is used for the time code generating unit 2A4 to 2D4 of the video cameras 2A to 2D has been described. However, the present invention is not restricted to this. Various other types can be widely applied as the time code generating unit 2A4 to 2D4, as long as they generate time codes synchronizing respectively based on the reference clock signal CLK.
As described above, according to the present invention, the editing device comprises the first displaying means displaying images based on a plurality of imaging results through converting to time lines with a standard of time codes added synchronizing with respective images on an identical time base respectively, the second displaying means displaying respective images corresponding to time codes designated by the time bases among the respective time lines displayed, the third displaying means displaying an image of a desired part sequentially designated, on the time bases, from the respective image displayed, as a clip respectively, wherein the third displaying means selects only the desired clip from clips corresponding to the respective imaging results configuring the desired parts designated. Thus, the temporal correlation between the clips extracted for every image of the desired part, which is designated sequentially from respective images on the time base can be easily understood. Thus, the editing system, in which the efficiency of editing work can be greatly improved, can be realized.
In addition, according to the present invention, the editing method comprises a first step displaying respective images based on a plurality of imaging results through converting to time lines with a standard of time codes added synchronizing with respective images on an identical time base respectively, a second step displaying respective images corresponding to time codes designated by the time bases among the respective time lines displayed, a third step displaying images of a desired part sequentially designated, on the time bases, from the respective displayed images as clips respectively, and a fourth step displaying only a clip selected from clips corresponding to the respective imaging results composing the designated desired parts. Thus, the temporal correlation between the clips extracted for every image of the desired part, which is designated sequentially from respective images on the time base can be easily understood. Thus, the editing system, in which the efficiency of editing work can be greatly improved, can be realized.
While there has been described in connection with the preferred embodiments of the invention, it will be obvious to those skilled in the art that various changes and modifications may be aimed, therefore, to cover in the appended claims all such changes and modifications as fall within the true spirit and scope of the invention.

Claims (6)

What is claimed is:
1. An editing device comprising:
image editing means for displaying image data from a plurality of sources on the basis of time codes of said image data on an identical time base;
first selecting means for selecting image data corresponding to a desired part of the time base;
clip displaying means for displaying selected image data as clip groups; each clip group being displayed in turn on alternating roll sides of a clip displaying area; and
second selecting means for sequentially selecting desired clips from said clip groups; each clip corresponding to image data from one of said plurality of sources.
2. The editing device according to claim 1, wherein each clip group comprises image data from all of said plurality of sources having time codes corresponding to the desired part of the time base.
3. The editing device according to claim 1, wherein said image data from said plurality of sources is information corresponding to video data of a scene taken by plural cameras from different angles.
4. The editing device according to claim 1, wherein said first selecting means selects an in-point and an out-point for each clip group.
5. The editing device according to claim 1, further comprising an edit list generating means for generating an edit list on the basis of the desired clips selected by said second selecting means.
6. The editing device according to claim 1, further comprising a monitor for displaying video data corresponding to the image data.
US09/587,438 1999-06-07 2000-06-05 Editing device and editing method Expired - Fee Related US6618058B1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP15975299A JP4009922B2 (en) 1999-06-07 1999-06-07 Editing apparatus and editing method
JPP11-159752 1999-06-07

Publications (1)

Publication Number Publication Date
US6618058B1 true US6618058B1 (en) 2003-09-09

Family

ID=15700497

Family Applications (1)

Application Number Title Priority Date Filing Date
US09/587,438 Expired - Fee Related US6618058B1 (en) 1999-06-07 2000-06-05 Editing device and editing method

Country Status (2)

Country Link
US (1) US6618058B1 (en)
JP (1) JP4009922B2 (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020054071A1 (en) * 2000-08-25 2002-05-09 Koichiro Yamaguchi Content editing apparatus, content editing method, content editing program, and computer-readable storage medium
US20020156881A1 (en) * 2000-10-03 2002-10-24 Klopp Lemon Ana H. Von HTTP transaction monitor with replay capacity
US20040066396A1 (en) * 2001-04-05 2004-04-08 Fujitsu Limited Information processing system, information processing method, medium and program
US20040105573A1 (en) * 2002-10-15 2004-06-03 Ulrich Neumann Augmented virtual environments
US20040225683A1 (en) * 2000-12-06 2004-11-11 Microsoft Corporation System and related methods for reducing source filter invocation in a development project
US20050120304A1 (en) * 2000-12-06 2005-06-02 Microsoft Corporation Interface and related methods for reducing source accesses in a development system
US20060168554A1 (en) * 2000-12-06 2006-07-27 Microsoft Corporation A System and Methods for Generating and Managing Filter Strings in a Filter Graph
US20060165354A1 (en) * 2005-01-26 2006-07-27 Samsung Electronics Co., Ltd. Digital video recording apparatus
US20060218299A1 (en) * 2000-12-06 2006-09-28 Microsoft Corporation Methods and Systems for Efficiently Processing Compressed and Uncompressed Media Content
US20060250522A1 (en) * 2005-05-04 2006-11-09 Pace Micro Technology Plc Television system
US20070098257A1 (en) * 2005-10-31 2007-05-03 Shesha Shah Method and mechanism for analyzing the color of a digital image
US20070097420A1 (en) * 2005-10-31 2007-05-03 Shesha Shah Method and mechanism for retrieving images
WO2008023352A2 (en) 2006-08-25 2008-02-28 Koninklijke Philips Electronics N.V. Method and apparatus for generating a summary
US20080109728A1 (en) * 2000-12-06 2008-05-08 Microsoft Corporation Methods and Systems for Effecting Video Transitions Represented By Bitmaps
US20090285565A1 (en) * 2008-05-15 2009-11-19 Sony Corporation Recording controlling device, recording controlling method, program used therein and recording device
US7673013B2 (en) 2000-12-06 2010-03-02 Microsoft Corporation Methods and systems for processing multi-media editing projects
US20100107080A1 (en) * 2008-10-23 2010-04-29 Motorola, Inc. Method and apparatus for creating short video clips of important events
FR2940481A1 (en) * 2008-12-23 2010-06-25 Thales Sa METHOD, DEVICE AND SYSTEM FOR EDITING ENRICHED MEDIA
US20100235857A1 (en) * 2007-06-12 2010-09-16 In Extenso Holdings Inc. Distributed synchronized video viewing and editing
US7853921B2 (en) 2000-12-06 2010-12-14 Microsoft Corporation Interface and related methods for dynamically generating a filter graph in a development system
US20150128195A1 (en) * 2011-12-29 2015-05-07 Sony Computer Entertainment Inc. Video reproduction system
KR101879166B1 (en) 2017-03-06 2018-07-23 (주)잼투고 A real-world studio system capable of producing contents using the control of a virtual studio and its operating method

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPWO2004021701A1 (en) 2002-08-27 2005-12-22 ソニー株式会社 Data processing apparatus and method, and program
KR20040033663A (en) * 2002-10-14 2004-04-28 삼성전자주식회사 Information storage medium having data structure for multi-angle and apparatus therefor

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5339393A (en) * 1993-04-15 1994-08-16 Sony Electronics, Inc. Graphical user interface for displaying available source material for editing
US5519828A (en) * 1991-08-02 1996-05-21 The Grass Valley Group Inc. Video editing operator interface for aligning timelines
US5613909A (en) * 1994-07-21 1997-03-25 Stelovsky; Jan Time-segmented multimedia game playing and authoring system
US5732184A (en) * 1995-10-20 1998-03-24 Digital Processing Systems, Inc. Video and audio cursor video editing system
US5760767A (en) * 1995-10-26 1998-06-02 Sony Corporation Method and apparatus for displaying in and out points during video editing
US5781188A (en) * 1996-06-27 1998-07-14 Softimage Indicating activeness of clips and applying effects to clips and tracks in a timeline of a multimedia work
US5786814A (en) * 1995-11-03 1998-07-28 Xerox Corporation Computer controlled display system activities using correlated graphical and timeline interfaces for controlling replay of temporal data representing collaborative activities
US5889519A (en) * 1996-03-26 1999-03-30 International Business Machines Corp. Method and system for a multimedia application development sequence editor using a wrap corral
US5892507A (en) * 1995-04-06 1999-04-06 Avid Technology, Inc. Computer system for authoring a multimedia composition using a visual representation of the multimedia composition
US5956453A (en) * 1996-04-12 1999-09-21 Hitachi Denshi Kabushiki Kaisha Method of editing moving image and apparatus of editing the same
US5999173A (en) * 1992-04-03 1999-12-07 Adobe Systems Incorporated Method and apparatus for video editing with video clip representations displayed along a time line
US6064379A (en) * 1996-06-24 2000-05-16 Sun Microsystems, Inc. System and method for synchronizing presentation of media stream playlists with real time
US6204840B1 (en) * 1997-04-08 2001-03-20 Mgi Software Corporation Non-timeline, non-linear digital multimedia composition method and system
US6353461B1 (en) * 1997-06-13 2002-03-05 Panavision, Inc. Multiple camera video assist control system
US6414686B1 (en) * 1998-12-01 2002-07-02 Eidos Plc Multimedia editing and composition system having temporal display

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5519828A (en) * 1991-08-02 1996-05-21 The Grass Valley Group Inc. Video editing operator interface for aligning timelines
US5999173A (en) * 1992-04-03 1999-12-07 Adobe Systems Incorporated Method and apparatus for video editing with video clip representations displayed along a time line
US5339393A (en) * 1993-04-15 1994-08-16 Sony Electronics, Inc. Graphical user interface for displaying available source material for editing
US5613909A (en) * 1994-07-21 1997-03-25 Stelovsky; Jan Time-segmented multimedia game playing and authoring system
US5892507A (en) * 1995-04-06 1999-04-06 Avid Technology, Inc. Computer system for authoring a multimedia composition using a visual representation of the multimedia composition
US5732184A (en) * 1995-10-20 1998-03-24 Digital Processing Systems, Inc. Video and audio cursor video editing system
US5760767A (en) * 1995-10-26 1998-06-02 Sony Corporation Method and apparatus for displaying in and out points during video editing
US5786814A (en) * 1995-11-03 1998-07-28 Xerox Corporation Computer controlled display system activities using correlated graphical and timeline interfaces for controlling replay of temporal data representing collaborative activities
US5889519A (en) * 1996-03-26 1999-03-30 International Business Machines Corp. Method and system for a multimedia application development sequence editor using a wrap corral
US5956453A (en) * 1996-04-12 1999-09-21 Hitachi Denshi Kabushiki Kaisha Method of editing moving image and apparatus of editing the same
US6064379A (en) * 1996-06-24 2000-05-16 Sun Microsystems, Inc. System and method for synchronizing presentation of media stream playlists with real time
US5781188A (en) * 1996-06-27 1998-07-14 Softimage Indicating activeness of clips and applying effects to clips and tracks in a timeline of a multimedia work
US6204840B1 (en) * 1997-04-08 2001-03-20 Mgi Software Corporation Non-timeline, non-linear digital multimedia composition method and system
US6353461B1 (en) * 1997-06-13 2002-03-05 Panavision, Inc. Multiple camera video assist control system
US6414686B1 (en) * 1998-12-01 2002-07-02 Eidos Plc Multimedia editing and composition system having temporal display

Cited By (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020054071A1 (en) * 2000-08-25 2002-05-09 Koichiro Yamaguchi Content editing apparatus, content editing method, content editing program, and computer-readable storage medium
US7240286B2 (en) * 2000-08-25 2007-07-03 Matsushita Electric Industrial Co., Ltd. Content editing apparatus, content editing method, content editing program, and computer-readable storage medium
US20020156881A1 (en) * 2000-10-03 2002-10-24 Klopp Lemon Ana H. Von HTTP transaction monitor with replay capacity
US20060218299A1 (en) * 2000-12-06 2006-09-28 Microsoft Corporation Methods and Systems for Efficiently Processing Compressed and Uncompressed Media Content
US8150954B2 (en) 2000-12-06 2012-04-03 Microsoft Corporation Methods and systems for processing multi-media editing projects
US20050120304A1 (en) * 2000-12-06 2005-06-02 Microsoft Corporation Interface and related methods for reducing source accesses in a development system
US20060168554A1 (en) * 2000-12-06 2006-07-27 Microsoft Corporation A System and Methods for Generating and Managing Filter Strings in a Filter Graph
US8010649B2 (en) 2000-12-06 2011-08-30 Microsoft Corporation Methods and systems for processing multi-media editing projects
US7581185B2 (en) * 2000-12-06 2009-08-25 Microsoft Corporation Interface and related methods for reducing source accesses in a development system
US7940275B2 (en) 2000-12-06 2011-05-10 Microsoft Corporation Interface and related methods for dynamically generating a filter graph in a development system
US20040225683A1 (en) * 2000-12-06 2004-11-11 Microsoft Corporation System and related methods for reducing source filter invocation in a development project
US7853921B2 (en) 2000-12-06 2010-12-14 Microsoft Corporation Interface and related methods for dynamically generating a filter graph in a development system
US7712106B2 (en) 2000-12-06 2010-05-04 Microsoft Corporation System and methods for generating and managing filter strings in a filter graph
US8612859B2 (en) 2000-12-06 2013-12-17 Microsoft Corporation Methods and systems for effecting video transitions represented by bitmaps
US7680898B2 (en) 2000-12-06 2010-03-16 Microsoft Corporation Systems for processing multi-media editing projects
US7673013B2 (en) 2000-12-06 2010-03-02 Microsoft Corporation Methods and systems for processing multi-media editing projects
US20080109728A1 (en) * 2000-12-06 2008-05-08 Microsoft Corporation Methods and Systems for Effecting Video Transitions Represented By Bitmaps
US7213089B2 (en) * 2001-04-05 2007-05-01 Fujitsu Limited System and method for displaying a preview indicating a time-based adjustment of image data or sound data
US20040066396A1 (en) * 2001-04-05 2004-04-08 Fujitsu Limited Information processing system, information processing method, medium and program
US7583275B2 (en) * 2002-10-15 2009-09-01 University Of Southern California Modeling and video projection for augmented virtual environments
US20040105573A1 (en) * 2002-10-15 2004-06-03 Ulrich Neumann Augmented virtual environments
US20060165354A1 (en) * 2005-01-26 2006-07-27 Samsung Electronics Co., Ltd. Digital video recording apparatus
US8233089B2 (en) * 2005-05-04 2012-07-31 Pace Plc. Television system
US20060250522A1 (en) * 2005-05-04 2006-11-09 Pace Micro Technology Plc Television system
US20070097420A1 (en) * 2005-10-31 2007-05-03 Shesha Shah Method and mechanism for retrieving images
US7831111B2 (en) * 2005-10-31 2010-11-09 Yahoo! Inc. Method and mechanism for retrieving images
US20070098257A1 (en) * 2005-10-31 2007-05-03 Shesha Shah Method and mechanism for analyzing the color of a digital image
US20100017716A1 (en) * 2006-08-25 2010-01-21 Koninklijke Philips Electronics N.V. Method and apparatus for generating a summary
WO2008023352A2 (en) 2006-08-25 2008-02-28 Koninklijke Philips Electronics N.V. Method and apparatus for generating a summary
CN101506892B (en) * 2006-08-25 2012-11-14 皇家飞利浦电子股份有限公司 Method and apparatus for generating a summary
WO2008023352A3 (en) * 2006-08-25 2008-04-24 Koninkl Philips Electronics Nv Method and apparatus for generating a summary
US8249153B2 (en) 2007-06-12 2012-08-21 In Extenso Holdings Inc. Distributed synchronized video viewing and editing
US20100235857A1 (en) * 2007-06-12 2010-09-16 In Extenso Holdings Inc. Distributed synchronized video viewing and editing
US8358914B2 (en) 2008-05-15 2013-01-22 Sony Corporation Recording controlling device, recording controlling method, program used therein and recording device
US20090285565A1 (en) * 2008-05-15 2009-11-19 Sony Corporation Recording controlling device, recording controlling method, program used therein and recording device
US20100107080A1 (en) * 2008-10-23 2010-04-29 Motorola, Inc. Method and apparatus for creating short video clips of important events
US9646648B2 (en) 2008-10-23 2017-05-09 Google Technology Holdings LLC Method and apparatus for creating short video clips of important events
US10424338B2 (en) 2008-10-23 2019-09-24 Google Technology Holdings LLC Method and apparatus for creating short video clips of important events
US10878849B2 (en) 2008-10-23 2020-12-29 Google Technology Holdings LLC Method and apparatus for creating short video clips of important events
FR2940481A1 (en) * 2008-12-23 2010-06-25 Thales Sa METHOD, DEVICE AND SYSTEM FOR EDITING ENRICHED MEDIA
WO2010072747A3 (en) * 2008-12-23 2010-09-16 Thales Method, device, and system for editing rich media
US20150128195A1 (en) * 2011-12-29 2015-05-07 Sony Computer Entertainment Inc. Video reproduction system
KR101879166B1 (en) 2017-03-06 2018-07-23 (주)잼투고 A real-world studio system capable of producing contents using the control of a virtual studio and its operating method

Also Published As

Publication number Publication date
JP4009922B2 (en) 2007-11-21
JP2000348469A (en) 2000-12-15

Similar Documents

Publication Publication Date Title
US6618058B1 (en) Editing device and editing method
EP1872268B1 (en) Icon bar display for video editing system
US7830570B2 (en) Device and method for edition of moving picture data
CN100366069C (en) Method of time shifting to simultaneously record and play data stream
US7804505B2 (en) Information processing apparatus and associated method of prioritizing content for playback
US7856145B2 (en) Image processing system and method therefor, image processing apparatus and method, image capturing apparatus and method, program recording medium, and program
US6560399B2 (en) Image recording and reproducing device and a medium storing an image recording and reproducing program
CN102917172B (en) Camera device, photograph control method and image reproduction method
US20040240005A1 (en) Editing apparatus and method using thumbnail image
US8224819B2 (en) Apparatus, method, and program for processing information
JPH08214251A (en) Apparatus and method for video effect generation
US20030219223A1 (en) Recording apparatus, editor terminal apparatus, recording medium, and video content editing support system and method using them
US6272279B1 (en) Editing method of moving images, editing apparatus and storage medium storing its editing method program
JP2005109923A (en) Video image recorder, editing data preparing device, and editing data preparing program
JPH11215458A (en) Image recorder and its method, image reproducing device and its method, image recording and reproducing device and its method, and recording medium for these
US20050232586A1 (en) Editing apparatus, editing method, program, and recording medium
JPH03291077A (en) Picture reproducing device
JP2006086612A (en) Information processor and information processing method, recording medium, and program
CN101325679B (en) Information processing apparatus, information processing method
JPH10276388A (en) Device, method for processing and reproducing image and recording medium
JP4016424B2 (en) Moving picture editing apparatus and moving picture editing method
WO2001037560A1 (en) Video searching method and video searching apparatus
JP2001292411A (en) Image data recorder and image data recording method
JPH1051734A (en) Dynamic image compiling device/method
JP4280975B2 (en) Data recording control device and method, data reproducing device and method, and program

Legal Events

Date Code Title Description
AS Assignment

Owner name: SONY CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:YASUI, SHIGEYA;REEL/FRAME:010877/0870

Effective date: 20000515

CC Certificate of correction
FPAY Fee payment

Year of fee payment: 4

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20110909