US20020175997A1 - Surveillance recording device and method - Google Patents

Surveillance recording device and method Download PDF

Info

Publication number
US20020175997A1
US20020175997A1 US10/153,440 US15344002A US2002175997A1 US 20020175997 A1 US20020175997 A1 US 20020175997A1 US 15344002 A US15344002 A US 15344002A US 2002175997 A1 US2002175997 A1 US 2002175997A1
Authority
US
United States
Prior art keywords
images
person
recording
essential
shot
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/153,440
Inventor
Yuji Takata
Shogo Hamasaki
Hideaki Matsuo
Kazuyuki Imagawa
Masafumi Yoshizawa
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Panasonic Holdings Corp
Original Assignee
Matsushita Electric Industrial Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Matsushita Electric Industrial Co Ltd filed Critical Matsushita Electric Industrial Co Ltd
Assigned to MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD. reassignment MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HAMASAKI, SHOGO, MATSUO, HIDEAKI, TAKATA, YUJI, YOSHIZAWA, MASAFUMI, IMAGAWA, KAZUYUKI
Publication of US20020175997A1 publication Critical patent/US20020175997A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/5838Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7837Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content
    • G06F16/784Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using objects detected or recognised in the video content the detected or recognised objects being people
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions

Definitions

  • the present invention relates to a surveillance recording device and method for surveilling the comings and goings of people with a camera.
  • Time lapse video has been used as one such conventional surveillance recording device which can record images over a long period of time.
  • Time lapse video is a device for compressing images obtained from a camera and storing the images onto a VHS video tape over a long period of time.
  • Another method for reducing tape usage, and consequent increased capacity includes compressing video images before recording using, for example, an image compression technique such as a conventional MPEG compression protocol.
  • the conventional surveillance recorder simply captures camera images. Therefore, after finishing recording, it is very difficult for an operator to search a target scene or person from long and massive image records.
  • an object of the invention is to provide a surveillance recording device and related techniques by which it becomes possible for an operator to easily search for a target person.
  • a surveillance recording device comprises cameras for shooting a target space, an image recording and reproducing unit for recording images shot by the cameras onto a recording medium and reproducing images from the recording medium, an essential image extracting unit for extracting essential images of a person from images shot by the cameras, and a retrieval information recording unit for recording retrieval information including the essential images.
  • facial images of people are included in the essential images.
  • a surveillance recording device In a surveillance recording device according to a third aspect of the invention, whole body images of people are included in the essential images.
  • a surveillance recording device comprises a personal characteristics detecting unit for detecting the personal characteristics based on the essential images extracted by the essential image extracting unit, and the retrieval information includes the personal characteristics.
  • personal characteristics include the heights of people.
  • a surveillance recording device comprises a best shot selecting unit for selecting a best shot among facial images of people, and the retrieval information includes the best shot facial image.
  • retrieval can be carried out by using the clearest facial images.
  • a surveillance recording device comprises a display unit and a display image generating unit for generating images to be displayed on the display unit, wherein the display image generating unit generates a thumbnail screen for displaying a list of essential images of people.
  • the display image generating unit generates a detailed information screen relating to a specific thumbnail specified on the thumbnail screen, and this detailed information screen includes essential images of a person, the personal characteristics, and the person shooting time.
  • the image recording and reproducing unit records images only in sections of a scene in which the essential image extracting unit has been able to extract essential images.
  • a surveillance recording device comprises at least cameras for stereoscopically shooting a target space, an image recording and reproducing unit for recording images shot by the cameras into a recording medium and reproducing images from this recording medium, a detection wall setting unit for setting a detection wall for detection of entry of people into the target space, and a collision detecting unit for detecting whether or not people collide with the detection wall, wherein the detection wall is a virtual wall composed of a plurality of voxels (three-dimensional volumes of space) depending on the positional relationship with the cameras, and the thickness of this detection wall is set to be sufficiently small with respect to the depth of the target space.
  • the essential image extracting unit extracts essential images of a person after the collision detecting unit detects collision of a person, and the retrieval information includes the time at which the collision detecting means detects collision of the person.
  • the image recording and reproducing unit starts recording images shot by the cameras after the collision detecting unit detects collision of a person.
  • FIG. 1 is a block diagram of a surveillance recording device according to an embodiment of the invention.
  • FIG. 2 is an explanatory view of a detection wall of the surveillance recording device of FIG. 1.
  • FIG. 3( a ) is an illustration of an image shot by cameras of the surveillance recording device of FIG. 1.
  • FIG. 3( b ) is an illustration of the facial image of the surveillance recording device of FIG. 1.
  • FIG. 3( c ) is an illustration of the whole body image of the surveillance recording device of FIG. 1.
  • FIG. 4 is an illustration of a template for facial direction judgment.
  • FIG. 5 is a flowchart of the surveillance recording device of FIG. 1.
  • FIG. 6 is a status transition drawing of a display screen of the surveillance recording device of FIG. 1.
  • FIG. 7( a ) is an illustration of a retrieval screen of the surveillance recording device of FIG. 1.
  • FIG. 7( b ) is an illustration of a thumbnail screen of the surveillance recording device of FIG. 1.
  • FIG. 7( c ) is an illustration of a detailed information screen of the surveillance recording device of FIG. 1.
  • the surveillance recording device includes a first camera 1 and a second camera 2 .
  • a first camera 1 and a second camera 2 Alternatively, if stereo vision is possible using a single stereo camera, only one stereo camera is sufficient. The number of cameras may be increased to three or more. Cameras 1 and 2 may be of types whose installation positions and parameters have been generally known. The positional relationship of the cameras 1 and 2 is described in detail later.
  • a control unit 4 controls the respective components shown in FIG. 1. Camera images shot by the first camera 1 and second camera 2 are inputted into the control unit 4 via an interface 3 .
  • a timer 5 supplies information including the current date and time to the control unit 4 .
  • An input unit 6 includes a keyboard and a mouse. The input unit 6 is used by an operator to input information such as detection wall information described later, recording start/end information, and retrieval information into the device.
  • a display unit 7 which may be an LCD or CRT, displays images required by an operator.
  • the images to be displayed on the display unit 7 are generated by a display image generating unit 8 in procedures described later.
  • An image recording and reproducing unit 9 reads and writes into a recording medium 90 , and stores and reproduces moving images.
  • the recording medium 90 is a large capacity digital recording and reproducing medium such as a DVD or DVC.
  • the image recording and reproducing unit 9 is a player for driving this medium.
  • operation time such as index search or fast-forwarding in reproduction of a recording medium
  • an analog medium such as a VHS may be used.
  • the recording format is optional, however, a format such as MPEG (Motion Picture Expert Group) coding in which images are compressed is desirable for recording over a long period of time without noticeable lowering in the apparent image quality.
  • MPEG Motion Picture Expert Group
  • a storing unit 10 is a memory or a hard disk, which is read and written by the control unit 4 .
  • Storing unit 10 stores information including detection wall information, facial images, whole body images, personal characteristics, and start/end times.
  • a detection wall setting unit 11 sets a detection wall described later.
  • a collision detecting unit 12 detects whether or not a person collides with the detection wall.
  • An essential image extracting unit 13 extracts essential images showing personal characteristics from images shot by the cameras 1 and 2 .
  • essential images are facial images and whole body images.
  • a personal characteristics detecting unit 14 detects personal characteristics.
  • the heights of people, calculated based on the whole body images, are used as personal characteristics.
  • the weights of people, as estimated from their heights, may be used as a personal characteristic.
  • Personal characteristics may also include gender, age bracket, body type, skin or hair color, and eye color.
  • a best shot selecting unit 15 selects best shots that most clearly show personal characteristics among essential images extracted by the essential image extracting unit 13 .
  • the best shot selecting unit 15 chooses an image showing a full face, and identifies this image as a best shot.
  • a facial image may be optionally selected if facial characteristics can be easily recognized in the image.
  • information stored in the storing unit 10 information such as facial images (best shots), whole body images, personal characteristics, and start/end time that can be utilized as indexes for retrieval of moving images within the storing unit 10 .
  • the indexes are later recorded in a database 17 as moving image retrieval information.
  • a database engine 16 retrieves the data from the database 17 or registers information into the database 17 under control of the control unit 4 .
  • the database 17 corresponds to the retrieval information recording means in claims hereof Moving image retrieval information may be directly recorded into a recording medium 90 without especially providing databases or database engines if the format of the recording medium 90 allows for such.
  • metadata and quoting data may, or may not, be in the same recording medium.
  • quoting data may be quoted via a network from a recording medium including the existence of metadata.
  • descriptors are used to categorize quoting data. While classifying retrieval information, pieces of data having a mutual relationship can be collectively and smartly quoted. This construction is also included in the present invention.
  • a detection wall may include an entry in one wall surface 21 of a target space 20 .
  • the entry includes doors 22 and 23 in a manner enabling them to be opened and closed.
  • the surveillance recording device of this embodiment surveilles movements of people and objects entering the target space 20 (in the direction of the arrow N).
  • the first camera 1 and second camera 2 are installed with their fields of view facing the wall surface 21 .
  • the relative positional relationship and parameters of the cameras are generally known.
  • a detection wall 24 (a virtual wall) is defined slightly in front of the doors 22 and 23 .
  • This detection wall 24 is a virtual thin wall parallel to the real wall surface 21 in this example,
  • the inside of the detection wall 24 is made up of a number of voxels 25 .
  • the detection wall 24 is as thin as possible in order to reduce the amount of detection processing.
  • the thickness of the detection wall is set to one voxel.
  • the thickness of the detection wall 24 may be set to be equivalent two or more voxels. However, at a minimum, the thickness is defined to be small with respect to the depth of the target space 20 (the length in the arrow N direction).
  • two cameras 1 and 2 are set so as to have points of view that are different from each other.
  • the cameras 1 and 2 shoot the wall surface 21 side from different directions.
  • Voxels that do not overlap the person are outside the person image in the camera image of at least one of the cameras 1 and 2 .
  • Voxels that overlap the person are inside the person images in camera images of all cameras.
  • the collision detecting unit 12 judges that the person has collided with the detection wall 24 .
  • the thin detection wall 24 composed of voxels, the fact that a person has entered the target space 20 can be detected. Furthermore, as mentioned above, by forming the detection wall 24 as thin as possible, the number of voxels to be examined by the collision detecting unit 12 is reduced. As a result, the amount of processing can be reduced and high-speed processing can be realized. The burden on system resources is correspondingly reduced.
  • Entry of a person can be detected using existing cameras for shooting surveillance images (installed in advance). Provision of other components, for example, an infrared-ray sensor for sensing passage of people in addition to the cameras, although permissible, is not necessary.
  • FIG. 2 shows a flat plane detection wall 24
  • the detection wall 24 since it is composed of virtual voxels, may be defined in any optional shape.
  • the detection wall can be freely changed into, for example, a curved shape, a shape with a bent portion, steps, or a shape enclosed by two or more surfaces in accordance with a target to be captured by surveillance.
  • the essential image extracting unit 13 extracts essential images (facial images and whole body images) showing personal characteristics among images shot by the cameras 1 and 2 .
  • a shot image is, for example, as shown in FIG. 3( a ) in which doors 22 and 23 are included in the background. An image of a woman is detected in front of the doors 22 and 23 to the left of the image.
  • the essential image extracting unit 13 uses two templates to define the essential image of the woman.
  • a first template T 1 of a small ellipse that is long horizontally is used for face detection.
  • the essential image extracting unit 13 carries on template matching in the usual manner to calculate the correlation between the shot image and these templates T 1 and T 2 , and calculates the point with maximum correlation in the shot image.
  • the essential image extracting unit 13 extracts images in the vicinity of the template T 1 as facial images. As shown in FIG. 3( c ), images in the vicinity of both templates T 1 and T 2 are extracted as whole body images.
  • the method for extracting faces from the shot image is not limited to the abovementioned method. Other than this, for example, a method involving detection of face parts and a method involving extraction of skin-color regions can be optionally selected.
  • the personal characteristics detecting unit 14 determines the height H of the whole body images extracted by the essential image extracting unit 13 as the height of a shot person as shown in FIG. 3( c ).
  • This height H can be easily determined from the number of voxels in the vertical direction of the whole body images since the geometric positions of the cameras 1 and 2 are known.
  • best shot selection by the best shot selecting unit 15 selects from several facial images the one which is most nearly a full facial image for determining the best shot. This shot is selected because the person face characteristics become most clear when the person turns his/her face frontward to directly face the cameras.
  • the best shot selecting unit 15 has a template of a standard full face as shown in FIG. 4, and carries out matching between the facial images extracted by the essential image extracting unit 13 and this template. Then, a facial image that is best matched with the template is regarded as the best shot.
  • the best shot selecting unit 15 determines an image with a maximum number of pixels within the skin color regions in a color space as a best shot.
  • a best shot can be determined by judging the timing.
  • the best shot selecting unit 15 may determine a best shot as the shot taken at this time.
  • step 1 the shooting and recording flow by the surveillance recording device according to the present embodiment begins in step 1 , wherein the control unit 4 clears the storing unit 10 , and the detection wall setting unit 11 sets the detection wall 24 (step 2 ).
  • the control unit 4 requires an operator to input detection wall information from the input unit 6 , or if the information has already been known, the information may be loaded from an external storing unit.
  • step 3 the control unit 4 starts inputting images from the first camera 1 and second camera 2 . Then, the control unit 4 directs the collision detecting unit 12 to detect whether or not a person has collided with the detection wall 24 , and the collision detecting unit 12 feeds back detection results to the control unit 4 .
  • control unit 4 advances the process to step 16 , and confirms that there are no instructions to end recording inputted from the input unit 6 , which then returns the process to step 3 .
  • step 5 the control unit 4 obtains current date and time information from the timer 5 , and stores this date and time information as a start time into the storing unit 10 .
  • step 6 the control unit 4 transmits a shot image to the essential image extracting unit 13 and commands the unit to extract essential images. Receiving this command, the essential image extracting unit 13 attempts to extract facial images and whole body images from the shot image.
  • the essential image extracting unit 13 adds facial images and whole body images into the storing unit 10 (step 8 ), and notifies the control unit 4 of the successful completion of extraction. Receiving this notification, the control unit 4 instructs the image recording and reproducing unit 9 to record the shot image as moving images. As a result, moving images are stored in the recording medium 90 .
  • step 7 when extraction has failed (for example, when a person is outside the fields of view of the cameras), the control unit 4 checks whether or not the essential images have been stored in the storing unit 10 in step 10 .
  • step 11 current date and time information is obtained from the timer 5 , and stores this date and time information into the storing unit 10 as an end time.
  • step 12 the control unit 4 transmits the whole body images in the storing unit 10 to the personal characteristics detecting unit 14 , directs the unit to calculate height as a personal characteristic, and stores the calculation result into the storing unit 10 .
  • step 13 the control unit 4 directs the best shot selecting unit 15 to select a best shot and obtains a selection result.
  • control unit 4 registers useful information including a best shot facial image, whole body image, start time, end time, and personal characteristics (moving image retrieval information) for retrieval of moving images in the database 17 by using the database engine 16 in step 14 .
  • step 15 the control unit 4 clears the moving image retrieval information in information stored in the storing unit 10 , advances the process to step 16 , and prepares for the next processing.
  • step 10 when there is no essential image in the storing unit 10 , the control unit 4 judges that the collision with the detection wall 24 was not caused by a person but by some other object, and advances the process to step 16 , in preparation for the next processing.
  • the display image generating unit 8 generates three types of screens, that is, a retrieval screen (FIG. 7( a )), thumbnail screen (FIG. 7( b )), and detailed information screen (FIG. 7( c )) in accordance with the circumstances, and displays them on the display unit 7 .
  • the abovementioned moving image retrieval information (registered in the database 17 ) is inputted.
  • a date and height are inputted.
  • the input information may be properly changed.
  • control unit 4 directs the database engine 16 to retrieve a corresponding piece of moving image retrieval information.
  • the retrieval results are transmitted to the display image generating unit 8 .
  • the display image generating unit 8 prepares thumbnails from corresponding facial images (best shots) and displays a list of thumbnails as shown in FIG. 7( b ).
  • the control unit 4 informs the display image generating unit 8 of the desired thumbnail based on the information inputted through the input unit 6 .
  • the display image generating unit 8 displays a detailed information screen as shown in FIG. 7( c ).
  • a facial image (best shot), whole body image, start time, and personal characteristics (height) are retrieved and displayed.
  • a corresponding moving image of the recording medium 90 from this start time is displayed at the same time.
  • retrieval is carried out in the retrieval screen first.
  • the retrieval process is omitted, and whole data is displayed on the thumbnail screen and selection of a target person is made.
  • thumbnail images and shooting time are displayed together on the thumbnail screen.
  • other personal characteristics such as, for example, gender, age and the like may be displayed.

Abstract

A surveillance recording device using cameras extracts facial images and whole body images of a person from images shot by the cameras. A height is calculated from the whole body images. Retrieval information, including a facial image (best shot), is associated with images in a recording medium and recorded into a database. The recorded data are utilized as an index for later retrieval from the recording medium. Facial images are displayed in a list of thumbnails to make it easy to retrieve a target person on a thumbnail screen. The images are displayed together with a moving image of a target person.

Description

    BACKGROUND OF THE INVENTION
  • 1. Field of the Invention [0001]
  • The present invention relates to a surveillance recording device and method for surveilling the comings and goings of people with a camera. [0002]
  • 2. Description of the Related Art [0003]
  • In facilities having objects to be protected, such as, for example, a bank, a surveillance camera is set up to surveille the comings and goings of people. Time lapse video has been used as one such conventional surveillance recording device which can record images over a long period of time. Time lapse video is a device for compressing images obtained from a camera and storing the images onto a VHS video tape over a long period of time. [0004]
  • In this device, in order to reduce the amount of data to be recorded, images inputted from a camera are recorded at fixed frame intervals while skipping intermediate frames. The skipping of frames lowers the image quality. However, skipping frames permits recording onto a videotape for a relatively long period of time with a particular tape length, compared to continuous recording. [0005]
  • Another method for reducing tape usage, and consequent increased capacity includes compressing video images before recording using, for example, an image compression technique such as a conventional MPEG compression protocol. [0006]
  • In conventional time lapse video, it is difficult to view recorded images since the level of image quality is poor. There is a possibility that the identity of an entering person cannot be distinguished. Furthermore, because of skipping frames between recorded frames, a key scene may occur during the skipped period. This raises the possibility that the scene including the person who has entered may not be recorded. [0007]
  • In addition to the abovementioned problems, the conventional surveillance recorder simply captures camera images. Therefore, after finishing recording, it is very difficult for an operator to search a target scene or person from long and massive image records. [0008]
  • For example, when a number of visitors enter the facility, in order to search for an image of a target person from long and massive image records, an operator searches for him/her while viewing all recorded images. This work is so troublesome and tiresome that operator fatigue may cause the image of a particular person to be missed. [0009]
  • OBJECTS AND SUMMARY OF THE INVENTION
  • Therefore, an object of the invention is to provide a surveillance recording device and related techniques by which it becomes possible for an operator to easily search for a target person. [0010]
  • A surveillance recording device according to a first aspect of the invention comprises cameras for shooting a target space, an image recording and reproducing unit for recording images shot by the cameras onto a recording medium and reproducing images from the recording medium, an essential image extracting unit for extracting essential images of a person from images shot by the cameras, and a retrieval information recording unit for recording retrieval information including the essential images. [0011]
  • By this construction, an operator can easily search for a target image by utilizing retrieval information. [0012]
  • In a surveillance recording device according to a second aspect of the invention, facial images of people are included in the essential images. [0013]
  • By this construction, the operator can easily intuitively carry out retrieval while referring to facial images of people. [0014]
  • In a surveillance recording device according to a third aspect of the invention, whole body images of people are included in the essential images. [0015]
  • By this construction, an operator can easily carry out retrieval based on physical characteristics or clothing while referring to the whole body images of people. [0016]
  • A surveillance recording device according to a fourth aspect of the invention comprises a personal characteristics detecting unit for detecting the personal characteristics based on the essential images extracted by the essential image extracting unit, and the retrieval information includes the personal characteristics. [0017]
  • By this construction, an operator can easily carry out retrieval based on personal characteristics of people. [0018]
  • In a surveillance recording device according to a fifth aspect of the invention, personal characteristics include the heights of people. [0019]
  • By this construction, an operator can easily carry out retrieval based on the height of a person. [0020]
  • A surveillance recording device according to a sixth aspect of the invention comprises a best shot selecting unit for selecting a best shot among facial images of people, and the retrieval information includes the best shot facial image. [0021]
  • By this construction, retrieval can be carried out by using the clearest facial images. [0022]
  • A surveillance recording device according to a seventh aspect of the invention comprises a display unit and a display image generating unit for generating images to be displayed on the display unit, wherein the display image generating unit generates a thumbnail screen for displaying a list of essential images of people. [0023]
  • By this construction, an operator can easily narrow down a target person on the thumbnail screen. [0024]
  • In a surveillance recording device according to an eighth aspect of the invention, the display image generating unit generates a detailed information screen relating to a specific thumbnail specified on the thumbnail screen, and this detailed information screen includes essential images of a person, the personal characteristics, and the person shooting time. [0025]
  • By this construction, an operator can narrow down a target person on the facial image thumbnail screen and review detailed information relating to the person, whereby the operator can efficiently carry out retrieval. [0026]
  • In a surveillance recording device according to a ninth aspect of the invention, the image recording and reproducing unit records images only in sections of a scene in which the essential image extracting unit has been able to extract essential images. [0027]
  • By this construction, useless images which do not include a person, but only background scenes are not recorded, so that a recording medium can be efficiently used. [0028]
  • A surveillance recording device according to a tenth aspect of the invention comprises at least cameras for stereoscopically shooting a target space, an image recording and reproducing unit for recording images shot by the cameras into a recording medium and reproducing images from this recording medium, a detection wall setting unit for setting a detection wall for detection of entry of people into the target space, and a collision detecting unit for detecting whether or not people collide with the detection wall, wherein the detection wall is a virtual wall composed of a plurality of voxels (three-dimensional volumes of space) depending on the positional relationship with the cameras, and the thickness of this detection wall is set to be sufficiently small with respect to the depth of the target space. [0029]
  • By this construction, only important sections are surveilled, and the calculation amount is reduced, whereby a speed increase and a saving of system resources can be achieved at the same time. Furthermore, the entry of people can be detected using only the cameras that are already installed in the surveillance recording device in advance, so that additional equipment such as a special sensor is unnecessary. [0030]
  • In a surveillance recording device according to an eleventh aspect of the invention, the essential image extracting unit extracts essential images of a person after the collision detecting unit detects collision of a person, and the retrieval information includes the time at which the collision detecting means detects collision of the person. [0031]
  • By this construction, useless extraction processes that would otherwise be performed until the person collides with the detection wall is eliminated, an operator can easily retrieve images, using the time of detection as a key. [0032]
  • In a surveillance recording device according to a twelfth aspect of the invention, the image recording and reproducing unit starts recording images shot by the cameras after the collision detecting unit detects collision of a person. [0033]
  • By this construction, useless image recording before a person collides with the detection wall is eliminated, the capacity of a recording medium can be efficiently used, and the time for seeking within the recording medium during retrieval can be reduced. [0034]
  • The above, and other objects, features and advantages of the present invention will become apparent from the following description read in conjunction with the accompanying drawings, in which like reference numerals designate the same elements.[0035]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a block diagram of a surveillance recording device according to an embodiment of the invention. [0036]
  • FIG. 2 is an explanatory view of a detection wall of the surveillance recording device of FIG. 1. [0037]
  • FIG. 3([0038] a) is an illustration of an image shot by cameras of the surveillance recording device of FIG. 1.
  • FIG. 3([0039] b) is an illustration of the facial image of the surveillance recording device of FIG. 1.
  • FIG. 3([0040] c) is an illustration of the whole body image of the surveillance recording device of FIG. 1.
  • FIG. 4 is an illustration of a template for facial direction judgment. [0041]
  • FIG. 5 is a flowchart of the surveillance recording device of FIG. 1. [0042]
  • FIG. 6 is a status transition drawing of a display screen of the surveillance recording device of FIG. 1. [0043]
  • FIG. 7([0044] a) is an illustration of a retrieval screen of the surveillance recording device of FIG. 1.
  • FIG. 7([0045] b) is an illustration of a thumbnail screen of the surveillance recording device of FIG. 1.
  • FIG. 7([0046] c) is an illustration of a detailed information screen of the surveillance recording device of FIG. 1.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
  • Referring to FIG. 1, the surveillance recording device according to the invention includes a [0047] first camera 1 and a second camera 2. Alternatively, if stereo vision is possible using a single stereo camera, only one stereo camera is sufficient. The number of cameras may be increased to three or more. Cameras 1 and 2 may be of types whose installation positions and parameters have been generally known. The positional relationship of the cameras 1 and 2 is described in detail later.
  • A [0048] control unit 4 controls the respective components shown in FIG. 1. Camera images shot by the first camera 1 and second camera 2 are inputted into the control unit 4 via an interface 3.
  • A [0049] timer 5 supplies information including the current date and time to the control unit 4. An input unit 6 includes a keyboard and a mouse. The input unit 6 is used by an operator to input information such as detection wall information described later, recording start/end information, and retrieval information into the device.
  • A [0050] display unit 7, which may be an LCD or CRT, displays images required by an operator. The images to be displayed on the display unit 7 are generated by a display image generating unit 8 in procedures described later.
  • An image recording and reproducing [0051] unit 9 reads and writes into a recording medium 90, and stores and reproduces moving images. Typically, the recording medium 90 is a large capacity digital recording and reproducing medium such as a DVD or DVC. The image recording and reproducing unit 9 is a player for driving this medium. Considering the operation time such as index search or fast-forwarding in reproduction of a recording medium, use of such a large capacity digital image recording and reproducing medium is advantageous. However, if operation time is not regarded as important, an analog medium such as a VHS may be used. The recording format is optional, however, a format such as MPEG (Motion Picture Expert Group) coding in which images are compressed is desirable for recording over a long period of time without noticeable lowering in the apparent image quality.
  • A storing [0052] unit 10 is a memory or a hard disk, which is read and written by the control unit 4. Storing unit 10 stores information including detection wall information, facial images, whole body images, personal characteristics, and start/end times.
  • A detection [0053] wall setting unit 11 sets a detection wall described later. A collision detecting unit 12 detects whether or not a person collides with the detection wall.
  • An essential [0054] image extracting unit 13 extracts essential images showing personal characteristics from images shot by the cameras 1 and 2. In this embodiment, essential images are facial images and whole body images.
  • A personal [0055] characteristics detecting unit 14 detects personal characteristics. In this embodiment, the heights of people, calculated based on the whole body images, are used as personal characteristics. The weights of people, as estimated from their heights, may be used as a personal characteristic. Personal characteristics may also include gender, age bracket, body type, skin or hair color, and eye color.
  • A best [0056] shot selecting unit 15 selects best shots that most clearly show personal characteristics among essential images extracted by the essential image extracting unit 13. In this embodiment, when several facial images of a subject are available, the best shot selecting unit 15 chooses an image showing a full face, and identifies this image as a best shot. As such a best shot, a facial image may be optionally selected if facial characteristics can be easily recognized in the image.
  • In information stored in the storing [0057] unit 10, information such as facial images (best shots), whole body images, personal characteristics, and start/end time that can be utilized as indexes for retrieval of moving images within the storing unit 10. The indexes are later recorded in a database 17 as moving image retrieval information. A database engine 16 retrieves the data from the database 17 or registers information into the database 17 under control of the control unit 4.
  • The [0058] database 17 corresponds to the retrieval information recording means in claims hereof Moving image retrieval information may be directly recorded into a recording medium 90 without especially providing databases or database engines if the format of the recording medium 90 allows for such.
  • In this case, the “recording medium” and “retrieval information recording means” in claims hereof are integrated, and this construction is also included in the present invention. [0059]
  • As above, two constructions, that is, a construction in which a “recording medium” and a “retrieval information recording means” in claims hereof are integrated together and a construction in which a “recording medium” and “retrieval information recording means” in claims hereof are separated from each other are described. Whichever construction is employed, typically, by using the format of MPEG7, retrieval information, moving images and still images shot by the [0060] cameras 1 and 2 or other necessary data (hereinafter, referred to as “quoting data” may be quoted in metadata (the format may be binary or ASCII).
  • In this case, metadata and quoting data may, or may not, be in the same recording medium. For example, quoting data may be quoted via a network from a recording medium including the existence of metadata. [0061]
  • Herein, by using the format of MPEG7, in metadata, descriptors are used to categorize quoting data. While classifying retrieval information, pieces of data having a mutual relationship can be collectively and smartly quoted. This construction is also included in the present invention. [0062]
  • Referring now to a detection wall may include an entry in one [0063] wall surface 21 of a target space 20. The entry includes doors 22 and 23 in a manner enabling them to be opened and closed. The surveillance recording device of this embodiment surveilles movements of people and objects entering the target space 20 (in the direction of the arrow N).
  • The [0064] first camera 1 and second camera 2 are installed with their fields of view facing the wall surface 21. The relative positional relationship and parameters of the cameras are generally known.
  • A detection wall [0065] 24 (a virtual wall) is defined slightly in front of the doors 22 and 23. This detection wall 24 is a virtual thin wall parallel to the real wall surface 21 in this example, The inside of the detection wall 24 is made up of a number of voxels 25. Preferably, the detection wall 24 is as thin as possible in order to reduce the amount of detection processing. For example, as shown in the figure, the thickness of the detection wall is set to one voxel. The thickness of the detection wall 24 may be set to be equivalent two or more voxels. However, at a minimum, the thickness is defined to be small with respect to the depth of the target space 20 (the length in the arrow N direction).
  • As mentioned above, in the present embodiment, two [0066] cameras 1 and 2 are set so as to have points of view that are different from each other. The cameras 1 and 2 shoot the wall surface 21 side from different directions.
  • When a person enters the inside of the [0067] target space 20 from the entry in the wall surface 21, a silhouette of the person is detected in the image planes of the respective cameras 1 and 2. When the person advances inside the target space 20 and collides with the detection wall 24, this collision is detected by the following procedures. The detection wall 24 is a virtual wall. Therefore, even when the person collides with the detection wall, he/she is not obstructed from advancing at all, does not recognize the collision, and passes through the detection wall 24.
  • Whether or not the voxels composing the [0068] detection wall 24 overlap the person is determined in accordance with the following principles.
  • Voxels that do not overlap the person are outside the person image in the camera image of at least one of the [0069] cameras 1 and 2. Voxels that overlap the person are inside the person images in camera images of all cameras.
  • In other words, if a certain voxel is within person images in camera images of all cameras, then that voxel overlaps the person. On the contrary, if a voxel is outside a person image in a camera image of either of the cameras, this means that the voxel does not overlap the person. [0070]
  • Therefore, if among the [0071] voxels 25 composing the detection wall 24, the number of voxels that are within person images in images of all cameras 1 and 2 is one or more, the collision detecting unit 12 judges that the person has collided with the detection wall 24.
  • On the contrary, if among the [0072] voxels 25 composing the detection wall 24, there is no voxel that is within person images in images of all cameras 1 and 2, it is judged that the person has not collided with the detection wall 24.
  • Thus, by means of the [0073] thin detection wall 24 composed of voxels, the fact that a person has entered the target space 20 can be detected. Furthermore, as mentioned above, by forming the detection wall 24 as thin as possible, the number of voxels to be examined by the collision detecting unit 12 is reduced. As a result, the amount of processing can be reduced and high-speed processing can be realized. The burden on system resources is correspondingly reduced.
  • Entry of a person can be detected using existing cameras for shooting surveillance images (installed in advance). Provision of other components, for example, an infrared-ray sensor for sensing passage of people in addition to the cameras, although permissible, is not necessary. [0074]
  • Although FIG. 2 shows a flat [0075] plane detection wall 24, the detection wall 24, since it is composed of virtual voxels, may be defined in any optional shape. The detection wall can be freely changed into, for example, a curved shape, a shape with a bent portion, steps, or a shape enclosed by two or more surfaces in accordance with a target to be captured by surveillance.
  • Incidentally, enclosure of a target to be captured by such a free encircling net is very difficult when using the abovementioned infrared ray sensor. [0076]
  • Referring now to FIG. 3, the essential [0077] image extracting unit 13 extracts essential images (facial images and whole body images) showing personal characteristics among images shot by the cameras 1 and 2.
  • A shot image is, for example, as shown in FIG. 3([0078] a) in which doors 22 and 23 are included in the background. An image of a woman is detected in front of the doors 22 and 23 to the left of the image.
  • The essential [0079] image extracting unit 13 uses two templates to define the essential image of the woman. A first template T1 of a small ellipse that is long horizontally is used for face detection. A second template T2 of a large ellipse that is long vertically, is used for detection of portions other than the face. The essential image extracting unit 13 carries on template matching in the usual manner to calculate the correlation between the shot image and these templates T1 and T2, and calculates the point with maximum correlation in the shot image.
  • Referring now to FIGS. [0080] 3(a) and 3(b), as a result, when a sufficiently good match is obtained (comparison with threshold values may be properly made), the essential image extracting unit 13 extracts images in the vicinity of the template T1 as facial images. As shown in FIG. 3(c), images in the vicinity of both templates T1 and T2 are extracted as whole body images.
  • As essential images, facial images alone are usually sufficient in practical use. The method for extracting faces from the shot image is not limited to the abovementioned method. Other than this, for example, a method involving detection of face parts and a method involving extraction of skin-color regions can be optionally selected. [0081]
  • As shown in FIG. 3([0082] c), the personal characteristics detecting unit 14 determines the height H of the whole body images extracted by the essential image extracting unit 13 as the height of a shot person as shown in FIG. 3(c). This height H can be easily determined from the number of voxels in the vertical direction of the whole body images since the geometric positions of the cameras 1 and 2 are known.
  • Referring now to FIG. 4, best shot selection by the best [0083] shot selecting unit 15 selects from several facial images the one which is most nearly a full facial image for determining the best shot. This shot is selected because the person face characteristics become most clear when the person turns his/her face frontward to directly face the cameras.
  • As described later, from the time a person collides with the [0084] detection wall 24 until the end of shoot of the person, a certain period of time elapses normally. Therefore, during the period, images of several frames are shot. In these several frames it is possible that several facial images of the person are obtained. The best shot selecting unit 15 selects an image in which the person is most clearly shot among these images.
  • Referring to FIG. 4, in the present embodiment, judgment of face direction is made. Concretely, the best [0085] shot selecting unit 15 has a template of a standard full face as shown in FIG. 4, and carries out matching between the facial images extracted by the essential image extracting unit 13 and this template. Then, a facial image that is best matched with the template is regarded as the best shot.
  • As another judgment of face direction, it is also allowed that the best [0086] shot selecting unit 15 determines an image with a maximum number of pixels within the skin color regions in a color space as a best shot.
  • Or, in place of the face direction judgment, a best shot can be determined by judging the timing. Herein, since the walking speed of a person can be ordinarily known, the time until the person is most clearly shot by the [0087] cameras 1 and 2 after a person collides with the detection wall 24 can be roughly estimated. Therefore, the best shot selecting unit 15 may determine a best shot as the shot taken at this time.
  • Referring now to FIG. 5, the shooting and recording flow by the surveillance recording device according to the present embodiment begins in [0088] step 1, wherein the control unit 4 clears the storing unit 10, and the detection wall setting unit 11 sets the detection wall 24 (step 2). Herein, the control unit 4 requires an operator to input detection wall information from the input unit 6, or if the information has already been known, the information may be loaded from an external storing unit.
  • Next, in [0089] step 3, the control unit 4 starts inputting images from the first camera 1 and second camera 2. Then, the control unit 4 directs the collision detecting unit 12 to detect whether or not a person has collided with the detection wall 24, and the collision detecting unit 12 feeds back detection results to the control unit 4.
  • If collision is not detected, the [0090] control unit 4 advances the process to step 16, and confirms that there are no instructions to end recording inputted from the input unit 6, which then returns the process to step 3.
  • When collision is detected, in [0091] step 5, the control unit 4 obtains current date and time information from the timer 5, and stores this date and time information as a start time into the storing unit 10.
  • Next, in [0092] step 6, the control unit 4 transmits a shot image to the essential image extracting unit 13 and commands the unit to extract essential images. Receiving this command, the essential image extracting unit 13 attempts to extract facial images and whole body images from the shot image.
  • At this point, when extraction is successfully carried out (step [0093] 7), the essential image extracting unit 13 adds facial images and whole body images into the storing unit 10 (step 8), and notifies the control unit 4 of the successful completion of extraction. Receiving this notification, the control unit 4 instructs the image recording and reproducing unit 9 to record the shot image as moving images. As a result, moving images are stored in the recording medium 90.
  • On the other hand, in [0094] step 7, when extraction has failed (for example, when a person is outside the fields of view of the cameras), the control unit 4 checks whether or not the essential images have been stored in the storing unit 10 in step 10.
  • When these have been stored, the [0095] control unit 4 judges that shooting of a person has been completed, and executes the next processing. First, in step 11, current date and time information is obtained from the timer 5, and stores this date and time information into the storing unit 10 as an end time.
  • In [0096] step 12, the control unit 4 transmits the whole body images in the storing unit 10 to the personal characteristics detecting unit 14, directs the unit to calculate height as a personal characteristic, and stores the calculation result into the storing unit 10. In step 13, the control unit 4 directs the best shot selecting unit 15 to select a best shot and obtains a selection result.
  • When the abovementioned processing is ended, the [0097] control unit 4 registers useful information including a best shot facial image, whole body image, start time, end time, and personal characteristics (moving image retrieval information) for retrieval of moving images in the database 17 by using the database engine 16 in step 14.
  • After completing registration, in [0098] step 15, the control unit 4 clears the moving image retrieval information in information stored in the storing unit 10, advances the process to step 16, and prepares for the next processing.
  • In [0099] step 10, when there is no essential image in the storing unit 10, the control unit 4 judges that the collision with the detection wall 24 was not caused by a person but by some other object, and advances the process to step 16, in preparation for the next processing.
  • In the processes mentioned above, the order of [0100] steps 11 through 13 may be freely interchanged.
  • By this construction, it can be understood that moving images in only a period in which a person is shot by the cameras after collision with the detection wall are detected are recorded in the storing [0101] unit 10. That is, useless recording in a period in which no person is shot by the cameras is omitted, thereby providing efficient operation is possible. In addition, moving image retrieval information is stored in the database 17. By using this information as an index, only important scenes can be easily retrieved and reproduced.
  • Next, the retrieval flow of surveillance results is explained with reference to FIG. 6 and FIG. 7. First, as shown in FIG. 6, in this retrieval, the display [0102] image generating unit 8 generates three types of screens, that is, a retrieval screen (FIG. 7(a)), thumbnail screen (FIG. 7(b)), and detailed information screen (FIG. 7(c)) in accordance with the circumstances, and displays them on the display unit 7.
  • These screens are changed from one to another in response to an operator clicking each button by using the [0103] input unit 6 as shown in FIG. 6.
  • First, in the retrieval screen shown in FIG. 7([0104] a), the abovementioned moving image retrieval information (registered in the database 17) is inputted. In the example shown in the figure, a date and height are inputted. However are just one example, and the input information may be properly changed.
  • Then, when the moving image retrieval information is inputted and the retrieval start button is clicked, the [0105] control unit 4 directs the database engine 16 to retrieve a corresponding piece of moving image retrieval information. The retrieval results are transmitted to the display image generating unit 8.
  • Then, the display [0106] image generating unit 8 prepares thumbnails from corresponding facial images (best shots) and displays a list of thumbnails as shown in FIG. 7(b).
  • When there are many person candidates, and it is not possible to display all thumbnails at the same time, a next screen button and a previous screen button are displayed on the screen. Then, when the button is clicked, the remaining thumbnail images are listed and displayed. [0107]
  • An operator checks this list and searches data to be examined based on facial images and clicks the thumbnail which he/she wants to check. [0108]
  • Then, the [0109] control unit 4 informs the display image generating unit 8 of the desired thumbnail based on the information inputted through the input unit 6. Receiving this information, the display image generating unit 8 displays a detailed information screen as shown in FIG. 7(c). In this example, at this point, from moving image retrieval information corresponding to the desired thumbnail, a facial image (best shot), whole body image, start time, and personal characteristics (height) are retrieved and displayed. In addition, a corresponding moving image of the recording medium 90 from this start time is displayed at the same time.
  • The display patterns shown in the figures are illustrations, and they may be properly changed for easy observation. [0110]
  • Herein, in this example, retrieval is carried out in the retrieval screen first. However, when the amount of data registered in the [0111] database 17 is small, it is also allowed that the retrieval process is omitted, and whole data is displayed on the thumbnail screen and selection of a target person is made.
  • The thumbnail images and shooting time are displayed together on the thumbnail screen. However, in addition to the shooting time, other personal characteristics that can be registered into the [0112] database 17, such as, for example, gender, age and the like may be displayed.
  • As shown in the figure, when a moving image is simultaneously displayed, incidental circumstances such as the number of persons and characters who entered at the same time with a target person can be grasped. This makes it easier to identify accomplices. [0113]
  • Having described preferred embodiments of the invention with reference to the accompanying drawings, it is to be understood that the invention is not limited to those precise embodiments, and that various changes and modifications may be effected therein by one skilled in the art without departing from the scope or spirit of the invention as defined in the appended claims. [0114]

Claims (26)

What is claimed is:
1. A surveillance recording device comprising:
at least one camera for shooting a target space;
an image recording and reproducing means for recording images shot by said cameras in a recording medium, and reproducing images from said recording medium;
an essential image extracting means for extracting essential images of a person from said images shot by said cameras; and
a retrieval information recording means for recording retrieval information including said essential images.
2. The surveillance recording device according to claim 1, wherein said essential images include facial images of said person.
3. The surveillance recording device according to claim 1, wherein said essential images include whole body images of said person.
4. The surveillance recording device according to claim 1, further comprising:
a personal characteristics detecting means for detecting personal characteristics based on said essential images extracted by said essential image extracting means; and
said retrieval information includes said personal characteristics.
5. The surveillance recording device according to claim 4, wherein said personal characteristics include a height of said person.
6. The surveillance recording device according to claim 2, further comprising:
a best shot selecting means for selecting a best shot among facial images of said person; and
said retrieval information includes said best shot facial image.
7. The surveillance recording device according to claim 2, further comprising:
a display means;
a display image generating means for generating display images to be displayed by said display means; and
said display image generating means includes means for generating a thumbnail screen for displaying a list of essential images of people.
8. The surveillance recording device according to claim 7, wherein:
said display image generating means includes means for generating a detailed information screen relating to a specified thumbnail on said thumbnail screen;
and said detailed information screen includes essential images, characteristics, and shooting times of people.
9. The surveillance recording device according to claim 1, wherein said image recording and reproducing means is effective to record images only in sections in which said essential image extracting means can extract essential images of people onto said recording medium.
10. A surveillance recording device comprising:
at least one camera for shooting a target space;
an image recording and reproducing means for recording images shot by said cameras onto a recording medium and for reproducing images from said recording medium;
a detection wall setting means for defining a detection wall for detecting entry of people into a target space;
a collision detecting means for detecting whether or not a person has collided with said detection wall;
said detection wall is a virtual wall composed of a plurality of voxels depending on positional relationship of said cameras;
a thickness of said detection wall is set to be small with respect to a depth of said target space.
11. The surveillance recording device according to claim 10, wherein:
said essential image extracting means includes means for extracting essential images of a person only after said collision detecting means detects collision of a person; and
said retrieval information includes a time at which said collision detecting means detects collision of said person.
12. The surveillance recording device according to claim 10, wherein said image recording and reproducing means includes means for starting recording images shot by said cameras after said collision detecting means detects collision of a person.
13. A surveillance recording device comprising:
at least one camera for shooting a target space;
an image recording and reproducing means for recording images shot by said cameras onto a recording medium and for reproducing images from said recording medium;
an essential image extracting means for extracting essential images of an object from images shot by said cameras; and
a retrieval information recording means for recording retrieval information including said essential images.
14. A surveillance recording method in which a target space is shot by cameras and shot images are recorded onto a recording medium, comprising:
means for extracting essential images of an object from said images shot by said cameras; and
means for retrieving information including said essential images associated with said images shot by said cameras and recorded.
15. A surveillance recording method in which a target space is shot by cameras and said shot images are recorded onto a recording medium, comprising:
extracting essential images of a person from said images shot by said cameras, and retrieval information including said essential images associated with said images shot by said cameras and recorded.
16. The surveillance recording method according to claim 15, wherein said essential images include facial images of said person.
17. The surveillance recording method according to claim 15, wherein said essential images include whole body images of said person.
18. The surveillance recording method according to claim 15, further comprising detecting at least one personal characteristic based on essential images and said personal characteristic is included in said retrieval information.
19. The surveillance recording method according to claim 18, wherein said personal characteristic includes a height of said person.
20. The surveillance recording method according to claim 16, further comprising selecting a best shot among facial images of said person and including said best shot facial image in said retrieval information.
21. The surveillance recording method according to claim 16, further comprising displaying a thumbnail screen containing a list of essential images of people.
22. The surveillance recording method according to claim 21, further comprising displaying a detailed information screen including essential images of a person, personal characteristics, and person shooting times that relate to a specified thumbnail selected from said thumbnail screen.
23. The surveillance recording method according to claim 15, further comprising recording images onto said recording medium only in sections in which essential images of people have been extracted.
24. A surveillance recording method further comprising:
at least stereoscopically shooting a target space and recording shot images onto a recording medium;
defining a virtual detection wall;
said step of defining includes defining a virtual detection wall composed of a plurality of voxels depending on said positional relationship of said cameras;
assigning a thickness to said virtual detection wall that is small with respect to a depth of said target space; and
detecting an entry of a person into said target space by detecting whether or not said person has collided with said detection wall.
25. The surveillance recording device according to claim 24, further comprising:
starting extracting of essential images of a person after detecting that said person has collided with said detection wall; and
including a time at which said person collides with said detection wall in said retrieval information.
26. The surveillance recording method according to claim 24, further comprising starting recording of images shot by said cameras only after detecting that a person has collided with said detection wall.
US10/153,440 2001-05-22 2002-05-21 Surveillance recording device and method Abandoned US20020175997A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2001151827 2001-05-22
JP2001-151827 2001-05-22

Publications (1)

Publication Number Publication Date
US20020175997A1 true US20020175997A1 (en) 2002-11-28

Family

ID=18996596

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/153,440 Abandoned US20020175997A1 (en) 2001-05-22 2002-05-21 Surveillance recording device and method

Country Status (2)

Country Link
US (1) US20020175997A1 (en)
EP (1) EP1260934A3 (en)

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030193582A1 (en) * 2002-03-29 2003-10-16 Fuji Photo Film Co., Ltd. Method for storing an image, method and system for retrieving a registered image and method for performing image processing on a registered image
US20030206654A1 (en) * 2002-05-01 2003-11-06 Heng-Tun Teng Replacing method of an object in a dynamic image
US20050163346A1 (en) * 2003-12-03 2005-07-28 Safehouse International Limited Monitoring an output from a camera
US20060195199A1 (en) * 2003-10-21 2006-08-31 Masahiro Iwasaki Monitoring device
US20070269082A1 (en) * 2004-08-31 2007-11-22 Matsushita Electric Industrial Co., Ltd. Surveillance Recorder and Its Method
US20080131073A1 (en) * 2006-07-04 2008-06-05 Sony Corporation Information processing apparatus and method, and program
US20080131092A1 (en) * 2006-12-05 2008-06-05 Canon Kabushiki Kaisha Video display system, video display method, and computer-readable medium
US20090009598A1 (en) * 2005-02-01 2009-01-08 Matsushita Electric Industrial Co., Ltd. Monitor recording device
US20090015663A1 (en) * 2005-12-22 2009-01-15 Dietmar Doettling Method and system for configuring a monitoring device for monitoring a spatial area
US20090052747A1 (en) * 2004-11-16 2009-02-26 Matsushita Electric Industrial Co., Ltd. Face feature collator, face feature collating method, and program
US20090285546A1 (en) * 2008-05-19 2009-11-19 Hitachi, Ltd. Recording and reproducing apparatus and method thereof
US20100158487A1 (en) * 2008-12-24 2010-06-24 Kabushiki Kaisha Toshiba Authoring device and authoring method
US20100253778A1 (en) * 2009-04-03 2010-10-07 Hon Hai Precision Industry Co., Ltd. Media displaying system and method
US20100299353A1 (en) * 2007-09-12 2010-11-25 Japan Women's University Moving Image Data Checking System, Moving Image Database Creating Method, and Registering System and Program for Registering Moving Image Data in Moving Image Database
US20110019003A1 (en) * 2009-07-22 2011-01-27 Hitachi Kokusai Electric Inc. Surveillance image retrieval apparatus and surveillance system
US20110164149A1 (en) * 2009-04-10 2011-07-07 Yuji Sugisawa Object detection device, object detection system, integrated circuit for object detection, camera with object detection function, and object detection method
US20120039506A1 (en) * 2008-08-27 2012-02-16 European Aeronautic Defence And Space Company - Eads France Method for identifying an object in a video archive
JP2012168250A (en) * 2011-02-10 2012-09-06 Ims:Kk Photographic sticker ejection device and control method thereof
US20130293739A1 (en) * 2007-03-15 2013-11-07 Sony Corporation Information processing apparatus, imaging apparatus, image display control method and computer program
US9858485B2 (en) * 2015-05-27 2018-01-02 Fujifilm Corporation Image processing device, image processing method and recording medium
US20180114078A1 (en) * 2015-08-20 2018-04-26 JVC Kenwood Corporation Vehicle detection device, vehicle detection system, and vehicle detection method
US11100691B2 (en) 2014-03-31 2021-08-24 Nec Corporation Image processing system, image processing method and program, and device

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7359529B2 (en) * 2003-03-06 2008-04-15 Samsung Electronics Co., Ltd. Image-detectable monitoring system and method for using the same
JP4624933B2 (en) 2005-03-16 2011-02-02 富士フイルム株式会社 IMAGING DEVICE, IMAGING METHOD, ALBUM CREATION DEVICE, ALBUM CREATION METHOD, ALBUM CREATION SYSTEM, AND PROGRAM
JP4424396B2 (en) * 2007-09-03 2010-03-03 ソニー株式会社 Data processing apparatus and method, data processing program, and recording medium on which data processing program is recorded
US8532390B2 (en) 2010-07-28 2013-09-10 International Business Machines Corporation Semantic parsing of objects in video
US10424342B2 (en) 2010-07-28 2019-09-24 International Business Machines Corporation Facilitating people search in video surveillance
CN106485869A (en) * 2016-07-11 2017-03-08 苏州超盛智能科技有限公司 A kind of wireless intelligent home security system
CN106251545B (en) * 2016-08-08 2018-11-30 张衡 A kind of intelligent family monitoring system
WO2018161339A1 (en) * 2017-03-10 2018-09-13 深圳市博信诺达经贸咨询有限公司 Application method and apparatus for big data searching in security monitoring
US10810255B2 (en) * 2017-09-14 2020-10-20 Avigilon Corporation Method and system for interfacing with a user to facilitate an image search for a person-of-interest
CN111538861B (en) * 2020-04-22 2023-08-15 浙江大华技术股份有限公司 Method, device, equipment and medium for image retrieval based on monitoring video

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5539199A (en) * 1994-02-19 1996-07-23 Leuze Electronic Gmbh + Co. Method for detecting objects in a monitored area
US5615622A (en) * 1992-11-25 1997-04-01 American Engineering Corporation Security module
US5745126A (en) * 1995-03-31 1998-04-28 The Regents Of The University Of California Machine synthesis of a virtual video camera/image of a scene from multiple video cameras/images of the scene in accordance with a particular perspective on the scene, an object in the scene, or an event in the scene
US6154755A (en) * 1996-07-31 2000-11-28 Eastman Kodak Company Index imaging system
US6331852B1 (en) * 1999-01-08 2001-12-18 Ati International Srl Method and apparatus for providing a three dimensional object on live video
US6400890B1 (en) * 1997-05-16 2002-06-04 Hitachi, Ltd. Image retrieving method and apparatuses therefor
US6476812B1 (en) * 1998-12-09 2002-11-05 Sony Corporation Information processing system, information processing method, and supplying medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE4113992A1 (en) * 1991-04-29 1992-11-05 Ameling Walter Automatic three=dimensional monitoring of hazardous room - using three cameras calibrated to universal standard to relate points in room to those of screen display

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5615622A (en) * 1992-11-25 1997-04-01 American Engineering Corporation Security module
US5539199A (en) * 1994-02-19 1996-07-23 Leuze Electronic Gmbh + Co. Method for detecting objects in a monitored area
US5745126A (en) * 1995-03-31 1998-04-28 The Regents Of The University Of California Machine synthesis of a virtual video camera/image of a scene from multiple video cameras/images of the scene in accordance with a particular perspective on the scene, an object in the scene, or an event in the scene
US6154755A (en) * 1996-07-31 2000-11-28 Eastman Kodak Company Index imaging system
US6400890B1 (en) * 1997-05-16 2002-06-04 Hitachi, Ltd. Image retrieving method and apparatuses therefor
US6476812B1 (en) * 1998-12-09 2002-11-05 Sony Corporation Information processing system, information processing method, and supplying medium
US6331852B1 (en) * 1999-01-08 2001-12-18 Ati International Srl Method and apparatus for providing a three dimensional object on live video

Cited By (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030193582A1 (en) * 2002-03-29 2003-10-16 Fuji Photo Film Co., Ltd. Method for storing an image, method and system for retrieving a registered image and method for performing image processing on a registered image
US20030206654A1 (en) * 2002-05-01 2003-11-06 Heng-Tun Teng Replacing method of an object in a dynamic image
US20060195199A1 (en) * 2003-10-21 2006-08-31 Masahiro Iwasaki Monitoring device
US7995843B2 (en) * 2003-10-21 2011-08-09 Panasonic Corporation Monitoring device which monitors moving objects
US7664292B2 (en) * 2003-12-03 2010-02-16 Safehouse International, Inc. Monitoring an output from a camera
US20050163346A1 (en) * 2003-12-03 2005-07-28 Safehouse International Limited Monitoring an output from a camera
US20070269082A1 (en) * 2004-08-31 2007-11-22 Matsushita Electric Industrial Co., Ltd. Surveillance Recorder and Its Method
USRE45551E1 (en) 2004-08-31 2015-06-09 Panasonic Intellectual Property Management Co., Ltd. Surveillance recorder and its method
US8131022B2 (en) 2004-08-31 2012-03-06 Panasonic Corporation Surveillance recorder and its method
US8073206B2 (en) * 2004-11-16 2011-12-06 Panasonic Corporation Face feature collator, face feature collating method, and program
US20090052747A1 (en) * 2004-11-16 2009-02-26 Matsushita Electric Industrial Co., Ltd. Face feature collator, face feature collating method, and program
US20090009598A1 (en) * 2005-02-01 2009-01-08 Matsushita Electric Industrial Co., Ltd. Monitor recording device
US8355045B2 (en) * 2005-02-01 2013-01-15 Panasonic Corporation Monitor recording device
US9151446B2 (en) 2005-12-22 2015-10-06 Pilz Gmbh & Co. Kg Method and system for configuring a monitoring device for monitoring a spatial area
US20090015663A1 (en) * 2005-12-22 2009-01-15 Dietmar Doettling Method and system for configuring a monitoring device for monitoring a spatial area
US9695980B2 (en) 2005-12-22 2017-07-04 Pilz Gmbh & Co. Kg Method and system for configuring a monitoring device for monitoring a spatial area
US9014537B2 (en) * 2006-07-04 2015-04-21 Sony Corporation Information processing apparatus and method, and program
US9672411B2 (en) 2006-07-04 2017-06-06 Sony Corporation Information processing apparatus and method, and program
US20080131073A1 (en) * 2006-07-04 2008-06-05 Sony Corporation Information processing apparatus and method, and program
US20080131092A1 (en) * 2006-12-05 2008-06-05 Canon Kabushiki Kaisha Video display system, video display method, and computer-readable medium
US8948570B2 (en) * 2006-12-05 2015-02-03 Canon Kabushiki Kaisha Video display system, video display method, and computer-readable medium
US20130293739A1 (en) * 2007-03-15 2013-11-07 Sony Corporation Information processing apparatus, imaging apparatus, image display control method and computer program
US9143691B2 (en) * 2007-03-15 2015-09-22 Sony Corporation Apparatus, method, and computer-readable storage medium for displaying a first image and a second image corresponding to the first image
US20100299353A1 (en) * 2007-09-12 2010-11-25 Japan Women's University Moving Image Data Checking System, Moving Image Database Creating Method, and Registering System and Program for Registering Moving Image Data in Moving Image Database
US10176848B2 (en) 2008-05-19 2019-01-08 Maxell, Ltd. Recording and reproducing apparatus and method thereof
US9159368B2 (en) 2008-05-19 2015-10-13 Hitachi Maxell, Ltd. Recording and reproducing apparatus and method thereof
US11727960B2 (en) 2008-05-19 2023-08-15 Maxell, Ltd. Recording and reproducing apparatus and method thereof
US11094350B2 (en) 2008-05-19 2021-08-17 Maxell, Ltd. Recording and reproducing apparatus and method thereof
US20090285546A1 (en) * 2008-05-19 2009-11-19 Hitachi, Ltd. Recording and reproducing apparatus and method thereof
US10418069B2 (en) 2008-05-19 2019-09-17 Maxell, Ltd. Recording and reproducing apparatus and method thereof
US11948605B2 (en) 2008-05-19 2024-04-02 Maxell, Ltd. Recording and reproducing apparatus and method thereof
US8594373B2 (en) * 2008-08-27 2013-11-26 European Aeronautic Defence And Space Company-Eads France Method for identifying an object in a video archive
US20120039506A1 (en) * 2008-08-27 2012-02-16 European Aeronautic Defence And Space Company - Eads France Method for identifying an object in a video archive
US20100158487A1 (en) * 2008-12-24 2010-06-24 Kabushiki Kaisha Toshiba Authoring device and authoring method
US8184945B2 (en) 2008-12-24 2012-05-22 Kabushiki Kaisha Toshiba Authoring device and authoring method
US20100253778A1 (en) * 2009-04-03 2010-10-07 Hon Hai Precision Industry Co., Ltd. Media displaying system and method
US8508603B2 (en) * 2009-04-10 2013-08-13 Panasonic Corporation Object detection device, object detection system, integrated circuit for object detection, and object detection method
US20110164149A1 (en) * 2009-04-10 2011-07-07 Yuji Sugisawa Object detection device, object detection system, integrated circuit for object detection, camera with object detection function, and object detection method
US9342744B2 (en) * 2009-07-22 2016-05-17 Hitachi Kokusai Electric Inc. Surveillance image retrieval apparatus and surveillance system
US20110019003A1 (en) * 2009-07-22 2011-01-27 Hitachi Kokusai Electric Inc. Surveillance image retrieval apparatus and surveillance system
JP2012168250A (en) * 2011-02-10 2012-09-06 Ims:Kk Photographic sticker ejection device and control method thereof
US11100691B2 (en) 2014-03-31 2021-08-24 Nec Corporation Image processing system, image processing method and program, and device
US11798211B2 (en) 2014-03-31 2023-10-24 Nec Corporation Image processing system, image processing method and program, and device
US9858485B2 (en) * 2015-05-27 2018-01-02 Fujifilm Corporation Image processing device, image processing method and recording medium
US20180114078A1 (en) * 2015-08-20 2018-04-26 JVC Kenwood Corporation Vehicle detection device, vehicle detection system, and vehicle detection method

Also Published As

Publication number Publication date
EP1260934A2 (en) 2002-11-27
EP1260934A3 (en) 2004-04-14

Similar Documents

Publication Publication Date Title
US20020175997A1 (en) Surveillance recording device and method
US8614741B2 (en) Method and apparatus for intelligent and automatic sensor control using multimedia database system
JP2003046911A (en) Monitoring recording system and method therefor
US7203693B2 (en) Instantly indexed databases for multimedia content analysis and retrieval
JP5730518B2 (en) Specific person detection system and specific person detection method
US8971585B2 (en) Image processing apparatus for retrieving object from moving image and method thereof
US20090060294A1 (en) Human image retrieval system
US8270806B2 (en) Information processing apparatus and method of controlling same
JP4973188B2 (en) Video classification device, video classification program, video search device, and video search program
JP4168940B2 (en) Video display system
WO2013069605A1 (en) Similar image search system
KR101960667B1 (en) Suspect Tracking Apparatus and Method In Stored Images
JP2018005555A (en) Image processing device, information processing device and method, as well as program
US20100118205A1 (en) Information processing apparatus and method of controlling same
US20110096994A1 (en) Similar image retrieval system and similar image retrieval method
US20040249848A1 (en) Method and apparatus for intelligent and automatic alert management using multimedia database system
JPH0993588A (en) Moving image processing method
JP2002344946A (en) Monitoring system
JP5768265B2 (en) Similar image search system
WO2018163398A1 (en) Similar image search system
US20050251741A1 (en) Methods and apparatus for capturing images
JP6203188B2 (en) Similar image search device
JP6214762B2 (en) Image search system, search screen display method
WO1999005865A1 (en) Content-based video access
JP5826513B2 (en) Similar image search system

Legal Events

Date Code Title Description
AS Assignment

Owner name: MATSUSHITA ELECTRIC INDUSTRIAL CO., LTD., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TAKATA, YUJI;HAMASAKI, SHOGO;MATSUO, HIDEAKI;AND OTHERS;REEL/FRAME:013140/0850;SIGNING DATES FROM 20020612 TO 20020619

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION