US20060120604A1 - Method and apparatus for detecting multi-view faces - Google Patents

Method and apparatus for detecting multi-view faces Download PDF

Info

Publication number
US20060120604A1
US20060120604A1 US11/285,172 US28517205A US2006120604A1 US 20060120604 A1 US20060120604 A1 US 20060120604A1 US 28517205 A US28517205 A US 28517205A US 2006120604 A1 US2006120604 A1 US 2006120604A1
Authority
US
United States
Prior art keywords
face
mode
detection
view
axis
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11/285,172
Inventor
Jung-Bae Kim
Chan-Min Park
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Assigned to SAMSUNG ELECTRONICS CO., LTD. reassignment SAMSUNG ELECTRONICS CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KIM, JUNG-BAE, PARK, CHAN-MIN
Publication of US20060120604A1 publication Critical patent/US20060120604A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/60Rotation of a whole image or part thereof

Definitions

  • the present invention relates to face detection, and more particularly, to a method and apparatus for detecting multi-view faces, by which any one of faces by all of X-rotation, Y-rotation, and Z-rotation is efficiently detected.
  • Face detection technology is used in various applications such as human computer interfaces and video monitoring systems, and image searching using a face as well as face recognition and has been thus increasingly important.
  • DCM digital contents management
  • Frontal face detection is satisfactory in a limited application environment such as a face recognition system that recognizes only a frontal face using a fixed camera but is deficient to be used in a usual environment.
  • many photographs and moving images used in image browsing and searching based on a face may be non-frontal face images.
  • studies on multi-view face detection have been performed to develop technology of detecting multi-view faces including a frontal face and a non-frontal face.
  • Z-rotation faces a Z-rotation direction of a face using a router network, rotated an image so that the face positively stands, and attempted face detection using a frontal face detector to detect faces obtained by Z-rotation (hereinafter, referred to as “Z-rotation faces”) [H. Rowley, S. Baluja, and T Kanade, “Neural Network-Based Face Detection”, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 20, no. 1, pp. 23-28, Jan. 1998].
  • an error occurring in the router network is not compensated for by the face detector.
  • X-rotation faces faces obtained by X-rotation
  • Y-rotation faces faces obtained by Y-rotation
  • Li et al. rotated an input image in three Z-axis directions and applied a detector-pyramid for detecting a Y-rotation face to each of the rotation results to simultaneously detect Y-rotation faces and Z-rotation faces [S. Z. Li and Z. Q. Zhang, “FloatBoost Learning and Statistical Face Detection”, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 26, no. 9, pp. 1112-1123, September 2004].
  • This approach cannot detect X-rotation faces and can only partially detect Z-rotation faces.
  • the approach is inefficient in that the same detector-pyramid is applied to a non-face portion three times.
  • Jones and Viola made a Y-rotation face detector and a Z-rotation face detector separately and used one of detectors for different angles according to a result of a pose estimator calculating the direction of Y-rotation or Z-rotation [M. Jones and P. Violar, “Fast Multi-View Face Detection”, Proc. Computer Vision and Pattern Recognition, March 2003].
  • this approach cannot compensate for an error of the pose estimator like the approach of Rowley et al. and cannot detect X-rotation faces.
  • An aspect of the present invention provides a method and apparatus for detecting multi-view faces, by which faces obtainable from all of X-rotation, Y-rotation, and Z-rotation are detected and a pose estimator is not used before a multi-view face detector is used, thereby preventing an error of the pose estimator from occurring and performing efficient operations.
  • a method of detecting multi-view face including the operations of (a) sequentially attempting to detect from an input image two mode faces among a first mode face made by up and down rotation, a second mode face made by leaning a head to the left and right, and a third mode face made by left and right rotation, (b) attempting to detect the remaining mode face that is not detected in operation (a), and (c) determining that a face is detected from the input image when the remaining mode face is detected in operation (b), wherein operation (b) comprises (b-1) arranging face detectors for all directions in parallel, when face detection succeeds in one direction, performing face detection in the same direction using a more complex face detector, and when face detection fails in one direction, performing face detection in a different direction; and (b-2) independently and separately arranging the face detectors for all directions, when face detection succeeds in one direction, performing face detection in the same direction using a more complex face detector, and when face detection fails, determining that a face is
  • an apparatus of detecting multi-view face having a face detection module including a subwindow generator receiving an input image and generating a subwindow with respect to the input image, a first face searcher receiving the subwindow and determining whether a whole-view face exists in the subwindow, a second face searcher sequentially searching for two mode faces among a first mode face made by up and down rotation, a second mode face made by leaning a head to the left and right, and a third mode face made by left and right rotation when the first face searcher determines that the whole-view face exists in the subwindow, a third face searcher searching for the remaining mode face that is not searched for by the second face searcher, and a controller controlling the subwindow generator to generate a new subwindow when one of the first face searcher, the second face searcher, and the third face searcher does not detect a face.
  • a computer-readable storage medium encoded with processing instructions for causing a processor to execute the above-described method.
  • FIG. 1 shows an example of directions in which a human face is rotated using a three-dimensional coordinate axis
  • FIG. 2 shows an example of angles at which a human face is rotated around an X-axis
  • FIG. 3 shows an example of angles at which a human face is rotated around a Y-axis
  • FIG. 4A shows an example of angles at which a human face is rotated around a Z-axis
  • FIG. 4B shows another example of angles at which a human face is rotated around the Z-axis
  • FIG. 5 illustrates a procedure for reducing the number of face detectors necessary for learning in a first Z-rotation mode for a frontal-view face, according to an embodiment of the present invention
  • FIG. 6 illustrates a procedure for reducing the number of face detectors necessary for learning in the first Z-rotation mode for a left-view face, according to an embodiment of the present invention
  • FIG. 7 shows faces to be learned with respect to a frontal-view face in the first Z-rotation mode in an embodiment of the present invention
  • FIG. 8 illustrates a procedure for reducing the number of face detectors necessary for learning in a second Z-rotation mode for a frontal-view face, according to an embodiment of the present invention
  • FIG. 9 illustrates a procedure for reducing the number of face detectors necessary for learning in the second Z-rotation mode for a left-view face, according to an embodiment of the present invention
  • FIG. 10 shows faces to be learned with respect to a frontal-view face in the second Z-rotation mode in an embodiment of the present invention
  • FIG. 11 is a block diagram of an apparatus for detecting a face according to an embodiment of the present invention.
  • FIG. 12 is a block diagram of a face detection module according to the embodiment illustrated in FIG. 11 ;
  • FIGS. 13A through 13C illustrate face search methods according to an embodiment of the present invention
  • FIG. 14 illustrates a method of detecting a face by combining three face search methods, according to an embodiment of the present invention
  • FIGS. 15A and 15B are flowcharts of a method of detecting a face according to an embodiment of the present invention.
  • FIG. 16 is a different type of flowchart of the method according to the embodiment illustrated in FIGS. 15A and 15B .
  • These computer program instructions may also be stored in a computer usable or computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instruction means that implement the function specified in the flowchart block or blocks.
  • the computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
  • a human face may be rotated around, for example, three-dimensional coordinate axes, i.e., an X-axis, a Y-axis, and a Z-axis.
  • an up-view, a frontal-view, and a down-view may be defined.
  • a left-view, a frontal-view, and a right-view may be defined.
  • views may be discriminated by a leaning angle.
  • the face is leaned at intervals of 30 degrees.
  • Rotation angles available to people will be described with respect to each of the X-, Y-, and Z-axes.
  • FIG. 2 shows an example of rotation angles of a human face around the X-axis.
  • Rotation around the X-axis i.e., X-rotation is referred to as “nodding rotation” or “out-of-plane rotation”.
  • the X-rotation i.e., up-and-down nodding
  • an up-view face in a range of [20°, 50°] has a high occurrence frequency and can be detected using a method of detecting a frontal-view face.
  • An up-view face in a range of [50°, 80°] rarely occurs and does not show face elements well and may be thus excluded from detection.
  • Preferably, with respect to the X-rotation only a down-view face in a range of [ ⁇ 60°, ⁇ 20°] and a frontal-view face in a range of [ ⁇ 20°, 50°] are detected.
  • FIG. 3 shows an example of rotation angles of a human face around the Y-axis. Rotation around the Y-axis, i.e., Y-rotation is referred to as “out-of-plane rotation”.
  • the Y-rotation (left and right rotation) has a range of [ ⁇ 180°, 180°]. However, in a range of [180°, ⁇ 90°] and a range of [90°, 180°], the back of a head occupies more than a face. Accordingly, in an embodiment of the present invention, only a left-view face in a range of [ ⁇ 90°, ⁇ 20°], a frontal-view face in a range of [ ⁇ 20°, 20°], and a right-view face in a range of [20°, 90°] are detected with respect to the Y-rotation.
  • Z-rotation When a face is rotated around the Z-axis, Z-rotation (left and right leaning) has a range of [ ⁇ 180°, 180°].
  • the Z-rotation is referred to as “in-plane rotation”.
  • a face may be defined to lean at intervals of 30° and 45°, which is respectively illustrated in FIGS. 4A and 4B .
  • a mode illustrated in FIG. 4A is referred to as a first Z-rotation mode and a mode illustrated in FIG. 4B is referred to as a second Z-rotation mode.
  • a left-leaned face, an upright face, and a right-leaned face are defined.
  • Table 1 shows the ranges of rotation angles of a face to be detected according to an embodiment of the present invention.
  • TABLE 1 Division X-rotation Y-rotation Z-rotation Description Up-and-down Left and right Left and right nodding rotation leaning Rotatable angle [ ⁇ 60°, 80°] [ ⁇ 180°, 180°] [ ⁇ 180°, 180°] Detec- Basic [ ⁇ 60°, 50°] [ ⁇ 90°, 90°] [ ⁇ 45°, 45°] tion mode target Extension [ ⁇ 60°, 50°] [ ⁇ 90°, 90°] [ ⁇ 180°, 180°] mode
  • a face detection apparatus may detect a face using cascaded classifiers, each of which is trained with conventional appearance-based pattern recognition, i.e., an AdaBoost algorithm.
  • the AdaBoost algorithm is an efficient learning algorithm that configures a plurality of simple and fast weak classifiers in a form of a weighted sum, thereby producing a single strong classifier which is fast and has a high success rate.
  • a strong classifier for detecting a particular face pose is referred to as a “face detector”.
  • the face detector discriminates a face from a non-face in an input image using a plurality of face patterns that it has learned. Accordingly, it is necessary to determine face patterns to be learned.
  • first Z-rotation mode 12 face detectors are needed in the extension mode and three face detectors are needed in the basic mode.
  • second Z-rotation mode 8 face detectors are needed in the extension mode and two face detectors are needed in the basic mode.
  • the number of face detectors for learning can be reduced by using rotation or mirroring (changing left and right coordinates), which is illustrated in FIG. 5 .
  • 12 face detectors can be made using three face detectors with respect to a left-view face.
  • a right-view face can be obtained by mirroring the left-view face.
  • 8 face detectors can be made using two face detectors with respect to a left-view face.
  • a right-view face can be obtained by mirroring the left-view face.
  • Table 2 shows the number of face detectors needed in an embodiment of the present invention. TABLE 2 Number of Number of necessary face face detectors detectors to learn First mode Basic mode 18 10 Extension mode 72 10 Second mode Basic mode 12 6 Extension mode 48 6
  • FIG. 11 is a block diagram of a face detection apparatus 1100 according to an embodiment of the present invention.
  • the face detection apparatus 1100 includes an image sensing module 1120 , a face detection module 1140 , a first storage module 1160 , and a second storage module 1180 .
  • the image sensing module 1120 has an imaging function like a camera.
  • the image sensing module 1120 senses an image of an object and provides the image to the face detection module 1140 .
  • the first storage module 1160 stores images sensed by the image sensing module 1120 or images captured by a user and provides the stored images to the face detection module 1140 according to the user's request.
  • the face detection module 1140 detects a human face from an image received from the image sensing module 1120 or the first storage module 1160 .
  • the second storage module 1180 stores an image of the detected human face.
  • the image stored in the second storage module 1180 may be transmitted to a display apparatus 1182 , a face recognition apparatus 1184 , or other image processing apparatus through a wired/wireless network 1186 .
  • the first storage module 1160 and the second storage module 1180 may be implemented as different storage areas in a physically single storage medium or may be implemented as different storage media, respectively.
  • the storage areas for the respective first and second storage modules 1160 and 1180 may be defined by a software program.
  • module means, but is not limited to, a software or hardware component, such as a Field Programmable Gate Array (FPGA) or Application Specific Integrated Circuit (ASIC), which performs certain tasks.
  • a module may advantageously be configured to reside on the addressable storage medium and configured to execute on one or more processors.
  • a module may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.
  • components such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables.
  • the functionality provided for in the components and modules may be combined into fewer components and modules or further separated into additional components and modules.
  • FIG. 12 is a block diagram of an example of the face detection module 1140 illustrated in FIG. 11 .
  • the face detection module 1140 includes a controller 1142 , a subwindow generator 1144 , a first face searcher 1146 , a second face searcher 1148 , and a third face searcher 1150 .
  • the subwindow generator 1144 generates a subwindow for an input image received from the image sensing module 1120 or the first storage module 1160 .
  • the subwindow is a portion clipped out of the input image in a predetermined size. For example, when the input image has a size of 320 ⁇ 240 pixels, if an image of 24 ⁇ 24 pixels is clipped, the clipped image will be a subwindow of the input image.
  • the subwindow generator 1144 defines a minimum subwindow size and increases the length or width of a subwindow step by step starting from the minimum subwindow size. In other words, the subwindow generator 1144 sequentially provides the first face searcher 1146 with subwindows generated while increasing the size of the subwindow step by step.
  • the first face searcher 1146 , the second face searcher 1148 , and the third face searcher 1150 perform operations to detect a face from each subwindow generated by the subwindow generator 1144 .
  • the controller 1142 controls the operation of the subwindow generator 1144 according to whether a face is detected by the operations of the first through third face searchers 1146 through 1150 .
  • the first face searcher 1146 Upon receiving a subwindow from the subwindow generator 1144 , the first face searcher 1146 searches for a face in the subwindow using a predetermined algorithm. If a face is detected, the first face searcher 1146 transmits the subwindow to the second face searcher 1148 . However, if no face is detected, the controller 1142 controls the subwindow generator 1144 to generate and transmit a new subwindow to the first face searcher 1146 . The second face searcher 1148 searches for a face in the received subwindow using a predetermined algorithm. If a face is detected, the second face searcher 1148 transmits the subwindow to the third face searcher 1150 .
  • the controller 1142 controls the subwindow generator 1144 to generate and transmit a new subwindow to the first face searcher 1146 .
  • the third face searcher 1150 searches for a face in the received subwindow using a predetermined algorithm. If a face is detected, the third face searcher 1150 stores the subwindow in a separate storage area (not shown). After face search is completely performed on all subwindows of the image provided by the image sensing module 1120 or the first storage module 1160 , face detection information of the image is stored in the second storage module 1180 based on the stored subwindows. However, if no face is detected by the third face searcher 1150 , the controller 1142 controls the subwindow generator 1144 to generate and transmit a new subwindow to the first face searcher 1146 .
  • the algorithms respectively used by the first face searcher 1146 , the second face searcher 1148 , and the third face searcher 1150 to search for a face will be described with reference to FIGS. 13A through 13C .
  • FIG. 13A illustrates a conventional coarse-to-fine search algorithm
  • FIG. 13B illustrates a conventional simple-to-complex search algorithm
  • FIG. 13C illustrates a parallel-to-separated search algorithm according to an embodiment of the present invention.
  • a whole-view classifier is made at an initial stage of a cascaded classifier and then classifiers for gradually narrower angles are made.
  • the coarse-to-fine search algorithm is used, a non-face is quickly removed in early stages so that entire detection time can be reduced.
  • the whole-view classifier only searches for the shape of a face in a given subwindow using information that has been learned, regardless of the pose of the face.
  • face detectors for all directions are arranged in parallel up to, for example, K-th stages, and face detectors for respective different directions are independently and separately arranged starting from a (K+1)-th stage.
  • face detection fails in one direction
  • face detection fails in one direction
  • face detection is performed in a different direction.
  • face detection in one direction succeeds, a subsequent stage in the same direction is continued.
  • face detection fails, a non-face is immediately determined and the face detection is terminated.
  • the direction of a face in an input image is determined in an initial stage, and thereafter, a face or a non-face is determined only with respect to the direction. Accordingly, a face detector having high accuracy and fast speed can be implemented.
  • FIG. 14 When the algorithms illustrated in FIGS. 13A through 13C are combined, a multi-view face detector illustrated in FIG. 14 can be made.
  • each block is a face detector detecting a face in a direction written in the block.
  • An area denoted by “A” operates in the same manner as the left part, and thus a description thereof is omitted.
  • a downward arrow indicates a flow of the operation when a face detector succeeds in detecting a face.
  • a rightward arrow indicates a flow of the operation when a face detector fails in detecting a face.
  • the first face searcher 1146 upon receiving a subwindow from the subwindow generator 1144 , discriminates a face from a non-face using a whole-view face detector based on already learned information in stage 1 ⁇ 1 ′.
  • the first face searcher 1146 transmits the subwindow to the second face searcher 1148 .
  • the second face searcher 1148 performs stage 2 ⁇ 2 ′ and stage 3 ⁇ 4 .
  • stage 2 ⁇ 2 ′ a frontal-view face and a down-view face are grouped with respect to the X-rotation and face detection is performed based on the already learned information.
  • stage 3 ⁇ 4 an upright face, a left-leaned face, and a right-leaned face are grouped with respect to the Z-rotation and face detection is performed based on the already learned information.
  • Stage 1 ⁇ 1 ′, stage 2 ⁇ 2 ′, and stage 3 ⁇ 4 are performed using the coarse-to-fine search algorithm. Face detectors performing stage 1 ⁇ 1 ′, stage 2 ⁇ 2 ′, and stage 3 ⁇ 4 internally use the simple-to-complex search algorithm.
  • stage M faces in all directions are classified based on the already learned information.
  • stage K when face detection succeeds, a subsequent downward stage is performed and when face detection fails, the operation shifts to a right face detector.
  • stage K when face detection succeeds, a subsequent downward stage is performed, but when face detection fails, a non-face is determined and face detection on a current subwindow is terminated. Accordingly, only with respect to a subwindow reaching stage M, it is determined that a face is detected.
  • Stage 5 ⁇ K and stage K+1 ⁇ M are performed using the parallel-to-separated search algorithm.
  • face detectors performing stage 5 ⁇ K and stage K+1 ⁇ M internally use the simple-to-complex search algorithm.
  • FIGS. 15A and 15B are flowcharts of a method of detecting a face according to an embodiment of the present invention.
  • FIG. 16 is a different type of flowchart of the method according to the embodiment illustrated in FIGS. 15A and 15B .
  • stage 2 ⁇ 2 ′ shown in FIG. 14 is omitted, and a stage for detecting a whole-view face is referred to as stage 1 ⁇ 2 .
  • stage 1 ⁇ 2 a stage for detecting a whole-view face
  • W represents “whole-view”
  • U represents “upright”
  • L represents “30° left-leaned”
  • R represents “30° right-leaned”
  • f represents “fail” indicating that face detection has failed
  • s represents “succeed” indicating that face detection has succeeded
  • NF represents “non-face”.
  • an initial value for detecting a face in the subwindow is set in operation S 1504 .
  • the initial value includes parameters n, N 1 , N 2 , K, and M.
  • the parameter “n” indicates a stage in the face detection.
  • the parameter N 1 indicates a reference value for searching for a whole-view face.
  • the parameter N 2 indicates a reference value for searching for an upright-view face, a left leaned-view face, and a right leaned-view face defined with respect to the Z-rotation.
  • the parameter M indicates a reference value for searching for a frontal-view face, a left-view face, and a right-view face defined with respect to the Y-rotation.
  • the parameter K indicates a reference value for discriminating a stage for arranging face detectors separately from a stage for arranging face detectors in parallel in the parallel-to-separated search algorithm according to an embodiment of the present invention.
  • the first face searcher 1146 searches for a whole-view face in stage “n” in operation S 1506 , i.e., 1602 . If a whole-view face is not detected, it is determined that no face exists in the subwindow. If a whole-view face is detected, the parameter “n” is increased by 1 in operation S 1508 . It is determined whether the value of “n” is greater than the value of N 1 in operation S 1510 . If the value of “n” is not greater than the value of N 1 , the method goes back to operation S 1506 . Since the parameter N 1 is set to 2 in the embodiment of the present invention, the first face searcher 1146 performs the simple-to-complex search algorithm on the whole-view face two times ( 1602 ⁇ 1604 ).
  • the second face searcher 1148 searches for an upright-view face in stage “n” in operation S 1512 (i.e., 1606 ).
  • the coarse-to-fine search algorithm is used.
  • a left leaned-view face is searched for in the same stage “n” in operation S 1560 (i.e., 1608 ). If the left leaned-view face is not detected in operation S 1560 , a right leaned-view face is searched for in the same stage “n” in operation S 1570 (i.e., 1610 ). If the right leaned-view face is not detected in operation S 1570 , it is determined that no face exists in the current subwindow.
  • the value of “n” is increased by 1 in operation S 1514 , S 1562 , or S 1572 , respectively, and it is determined whether the increased value of “n” is greater than the value of N 2 in operation S 1516 , S 1564 , or S 1574 , respectively. If the value of “n” is not greater than the value of N 2 , the method goes back to operation S 1512 , S 1560 , or S 1570 .
  • the second face searcher 1148 performs the simple-to-complex search algorithm on the upright-view face, the left leaned-view face, or the right leaned-view face two times ( 1606 ⁇ 1612 , 1608 ⁇ 1614 , or 1610 ⁇ 1616 ).
  • the third face searcher 1150 searches for an upright frontal-view face in stage “n” in operation S 1520 (i.e., 1618 ). If the upright frontal-view face is not detected in operation S 1520 ( 1618 ), an upright left leaned-view face is detected in the same stage “n” in operation S 1526 (i.e., 1620 ). If the upright left leaned-view face is not detected in operation S 1526 , an upright right leaned-view face is searched for in the same stage “n” in operation S 1532 (i.e., 1622 ). If the upright right leaned-view face is not detected in operation S 1532 , face detection is continued in the 1-block (S 1566 ) or the 11-block (S 1576 ).
  • the value of “n” is increased by 1 in operation S 1522 , S 1528 , or S 1534 , respectively, and it is determined whether the increased value of “n” is greater than the value of K in operation S 1524 , S 1530 , or S 1535 , respectively. If the value of “n” is not greater than the value of K, the method goes back to operation S 1520 , S 1526 , or S 1532 .
  • the third face searcher 1150 performs the simple-to-complex search algorithm on the upright frontal-view face, the upright left leaned-view face, or the upright right leaned-view face up to a maximum of 6 times ( 1618 ⁇ 1624 , 1620 ⁇ 1626 , or 1622 ⁇ 1628 ).
  • the third face searcher 1150 searches for an upright frontal-view face in stage “n” in operation S 1540 (i.e., 1630 ). If the upright frontal-view face is not detected in operation S 1540 ( 1630 ), it is determined that no face exists in the current subwindow. If the upright frontal-view face is detected in operation S 1540 , the value of “n” is increased by 1 in operation S 1542 and it is determined whether the increased value of “n” is greater than the value of M in operation S 1544 . If the increased value of “n” is not greater than the value of M, the method goes back to operation S 1540 . If the increased value of “n” is greater than the value of M, it is determined that a face exists in the current subwindow.
  • the third face searcher 1150 operates using the parallel-to-separated search algorithm according to an embodiment of the present invention and the conventional simple-to-complex search algorithm.
  • face detectors for all directions are arranged in parallel up to stage K and are arranged separately from each other from stage K+1 to stage M, and the simple-to-complex search algorithm is used when a stage shifts.
  • X-rotation faces are detected first, Z-rotation faces are detected next, and Y-rotation faces are detected finally.
  • X-rotation faces are detected first, Z-rotation faces are detected next, and Y-rotation faces are detected finally.
  • such order is just an example, and it will be obvious to those skilled in the art that the order may be changed in face detection.
  • any one of the X-rotation faces, Y-rotation faces and Z-rotation faces can be detected.
  • a pose estimator is not used prior to a multi-view face detector, an error of the pose estimator does not occur and accuracy and operating speed increase. As a result, an efficient operation can be performed.
  • the above-described embodiments of the present invention can be used for any field requiring face recognition such as credit cards, cash cards, digital social security cards, cards needing identity authentication, terminal access control, control systems in public places, digital album, and recognition of photographs of criminals.
  • the present invention can also be used for security monitoring systems.

Abstract

A method and apparatus for detecting multi-view faces. The method of detecting multi-view face, includes the operations of (a) sequentially attempting to detect from an input image two mode faces among a first mode face made by up and down rotation, a second mode face made by leaning a head to the left and right, and a third mode face made by left and right rotation; (b) attempting to detect the remaining mode face that is not detected in operation (a); and (c) determining that a face is detected from the input image when the remaining mode face is detected in operation (b), wherein operation (b) comprises (b-1) arranging face detectors for all directions in parallel, when face detection succeeds in one direction, performing face detection in the same direction using a more complex face detector, and when face detection fails in one direction, performing face detection in a different direction; and (b-2) independently and separately arranging the face detectors for all directions, when face detection succeeds in one direction, performing face detection in the same direction using a more complex face detector, and when face detection fails, determining that a face is not detected from the input image.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application claims priority from Korean Patent Application No. 10-2004-0102411 filed on Dec. 7, 2004 in the Korean Intellectual Property Office, the disclosure of which is incorporated herein by reference.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to face detection, and more particularly, to a method and apparatus for detecting multi-view faces, by which any one of faces by all of X-rotation, Y-rotation, and Z-rotation is efficiently detected.
  • 2. Description of Related Art
  • Face detection technology is used in various applications such as human computer interfaces and video monitoring systems, and image searching using a face as well as face recognition and has been thus increasingly important.
  • In particular, in digital contents management (DCM) which is technology of browsing and searching photographs and video images to allow a user to easily obtain desired information from a huge amount of multimedia data, a method of detecting and recognizing a face is essential to classify a large amount of multimedia video by individuals.
  • In addition, with the improvement in the performance of mobile phone cameras and the calculation performance of mobile phones, the development of user authentication technology using face recognition has been demanded for mobile phones.
  • Many studies have been performed on face detection for recent several years, but they concentrate on only frontal face detection. Frontal face detection is satisfactory in a limited application environment such as a face recognition system that recognizes only a frontal face using a fixed camera but is deficient to be used in a usual environment. In particular, many photographs and moving images used in image browsing and searching based on a face may be non-frontal face images. Accordingly, studies on multi-view face detection have been performed to develop technology of detecting multi-view faces including a frontal face and a non-frontal face.
  • Rowley et al. calculated a Z-rotation direction of a face using a router network, rotated an image so that the face positively stands, and attempted face detection using a frontal face detector to detect faces obtained by Z-rotation (hereinafter, referred to as “Z-rotation faces”) [H. Rowley, S. Baluja, and T Kanade, “Neural Network-Based Face Detection”, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 20, no. 1, pp. 23-28, Jan. 1998]. However, an error occurring in the router network is not compensated for by the face detector. As a result, a detection rate is decreased and faces obtained by X-rotation (hereinafter, referred to as “X-rotation faces”) and faces obtained by Y-rotation (hereinafter, referred to as “Y-rotation faces”) cannot be detected.
  • Meanwhile, Schneiderman et al. detected Y-rotation faces using three independent profile detectors [H. Schneiderman and T. Kanade, “Object Detection Using the Statistics of Parts”, Int'l J. Computer Vision, vol. 56, no. 3, pp. 151-177, February 2004]. However, this approach requires three times longer detection time than the approach using a frontal face detector and cannot detect X-rotation faces and Z-rotation faces.
  • Li et al. rotated an input image in three Z-axis directions and applied a detector-pyramid for detecting a Y-rotation face to each of the rotation results to simultaneously detect Y-rotation faces and Z-rotation faces [S. Z. Li and Z. Q. Zhang, “FloatBoost Learning and Statistical Face Detection”, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 26, no. 9, pp. 1112-1123, September 2004]. This approach cannot detect X-rotation faces and can only partially detect Z-rotation faces. In addition, the approach is inefficient in that the same detector-pyramid is applied to a non-face portion three times.
  • Jones and Viola made a Y-rotation face detector and a Z-rotation face detector separately and used one of detectors for different angles according to a result of a pose estimator calculating the direction of Y-rotation or Z-rotation [M. Jones and P. Violar, “Fast Multi-View Face Detection”, Proc. Computer Vision and Pattern Recognition, March 2003]. However, this approach cannot compensate for an error of the pose estimator like the approach of Rowley et al. and cannot detect X-rotation faces.
  • Although various approaches for multi-view face detection have been proposed as described above, they are limited in performance. In other words, only a part of X-rotation faces, Y-rotation faces, and Z-rotation faces can be detected or an error of a pose estimator is not compensated for. Accordingly, a solution to limitation on performance is desired.
  • BRIEF SUMMARY
  • An aspect of the present invention provides a method and apparatus for detecting multi-view faces, by which faces obtainable from all of X-rotation, Y-rotation, and Z-rotation are detected and a pose estimator is not used before a multi-view face detector is used, thereby preventing an error of the pose estimator from occurring and performing efficient operations.
  • According to an aspect of the present invention, there is provided a method of detecting multi-view face, including the operations of (a) sequentially attempting to detect from an input image two mode faces among a first mode face made by up and down rotation, a second mode face made by leaning a head to the left and right, and a third mode face made by left and right rotation, (b) attempting to detect the remaining mode face that is not detected in operation (a), and (c) determining that a face is detected from the input image when the remaining mode face is detected in operation (b), wherein operation (b) comprises (b-1) arranging face detectors for all directions in parallel, when face detection succeeds in one direction, performing face detection in the same direction using a more complex face detector, and when face detection fails in one direction, performing face detection in a different direction; and (b-2) independently and separately arranging the face detectors for all directions, when face detection succeeds in one direction, performing face detection in the same direction using a more complex face detector, and when face detection fails, determining that a face is not detected from the input image.
  • According to another aspect of the present invention, there is provided an apparatus of detecting multi-view face having a face detection module including a subwindow generator receiving an input image and generating a subwindow with respect to the input image, a first face searcher receiving the subwindow and determining whether a whole-view face exists in the subwindow, a second face searcher sequentially searching for two mode faces among a first mode face made by up and down rotation, a second mode face made by leaning a head to the left and right, and a third mode face made by left and right rotation when the first face searcher determines that the whole-view face exists in the subwindow, a third face searcher searching for the remaining mode face that is not searched for by the second face searcher, and a controller controlling the subwindow generator to generate a new subwindow when one of the first face searcher, the second face searcher, and the third face searcher does not detect a face.
  • According to another aspect of the present invention, there is provided a computer-readable storage medium encoded with processing instructions for causing a processor to execute the above-described method.
  • Additional and/or other aspects and advantages of the present invention will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The above and/or other aspects and advantages of the present invention will become apparent and more readily appreciated from the following detailed description, taken in conjunction with the accompanying drawings of which:
  • FIG. 1 shows an example of directions in which a human face is rotated using a three-dimensional coordinate axis;
  • FIG. 2 shows an example of angles at which a human face is rotated around an X-axis;
  • FIG. 3 shows an example of angles at which a human face is rotated around a Y-axis;
  • FIG. 4A shows an example of angles at which a human face is rotated around a Z-axis;
  • FIG. 4B shows another example of angles at which a human face is rotated around the Z-axis;
  • FIG. 5 illustrates a procedure for reducing the number of face detectors necessary for learning in a first Z-rotation mode for a frontal-view face, according to an embodiment of the present invention;
  • FIG. 6 illustrates a procedure for reducing the number of face detectors necessary for learning in the first Z-rotation mode for a left-view face, according to an embodiment of the present invention;
  • FIG. 7 shows faces to be learned with respect to a frontal-view face in the first Z-rotation mode in an embodiment of the present invention;
  • FIG. 8 illustrates a procedure for reducing the number of face detectors necessary for learning in a second Z-rotation mode for a frontal-view face, according to an embodiment of the present invention;
  • FIG. 9 illustrates a procedure for reducing the number of face detectors necessary for learning in the second Z-rotation mode for a left-view face, according to an embodiment of the present invention;
  • FIG. 10 shows faces to be learned with respect to a frontal-view face in the second Z-rotation mode in an embodiment of the present invention;
  • FIG. 11 is a block diagram of an apparatus for detecting a face according to an embodiment of the present invention;
  • FIG. 12 is a block diagram of a face detection module according to the embodiment illustrated in FIG. 11;
  • FIGS. 13A through 13C illustrate face search methods according to an embodiment of the present invention;
  • FIG. 14 illustrates a method of detecting a face by combining three face search methods, according to an embodiment of the present invention;
  • FIGS. 15A and 15B are flowcharts of a method of detecting a face according to an embodiment of the present invention; and
  • FIG. 16 is a different type of flowchart of the method according to the embodiment illustrated in FIGS. 15A and 15B.
  • DETAILED DESCRIPTION OF EMBODIMENTS
  • Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like elements throughout. The embodiments are described below in order to explain the present invention by referring to the figures.
  • The present invention is described hereinafter with reference to flowchart illustrations of methods according to embodiments of the invention. It is to be understood that each block of the flowchart illustrations, and combinations of blocks in the flowchart illustrations, can be implemented by computer program instructions. These computer program instructions can be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart block or blocks. These computer program instructions may also be stored in a computer usable or computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer usable or computer-readable memory produce an article of manufacture including instruction means that implement the function specified in the flowchart block or blocks. The computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions that execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart block or blocks.
  • To detect multi-view faces, it is necessary to find and define face rotation angles available to people.
  • As shown in FIG. 1, a human face may be rotated around, for example, three-dimensional coordinate axes, i.e., an X-axis, a Y-axis, and a Z-axis.
  • When the face is rotated around the X-axis, an up-view, a frontal-view, and a down-view may be defined.
  • When the face is rotated around the Y-axis, a left-view, a frontal-view, and a right-view may be defined.
  • When the face is rotated around the Z-axis, views may be discriminated by a leaning angle. In FIG. 1, the face is leaned at intervals of 30 degrees.
  • Rotation angles available to people will be described with respect to each of the X-, Y-, and Z-axes.
  • FIG. 2 shows an example of rotation angles of a human face around the X-axis. Rotation around the X-axis, i.e., X-rotation is referred to as “nodding rotation” or “out-of-plane rotation”. The X-rotation (i.e., up-and-down nodding) has a range of about [−60°, 80°]. However, an up-view face in a range of [20°, 50°] has a high occurrence frequency and can be detected using a method of detecting a frontal-view face. An up-view face in a range of [50°, 80°] rarely occurs and does not show face elements well and may be thus excluded from detection. Preferably, with respect to the X-rotation, only a down-view face in a range of [−60°, −20°] and a frontal-view face in a range of [−20°, 50°] are detected.
  • FIG. 3 shows an example of rotation angles of a human face around the Y-axis. Rotation around the Y-axis, i.e., Y-rotation is referred to as “out-of-plane rotation”.
  • The Y-rotation (left and right rotation) has a range of [−180°, 180°]. However, in a range of [180°, −90°] and a range of [90°, 180°], the back of a head occupies more than a face. Accordingly, in an embodiment of the present invention, only a left-view face in a range of [−90°, −20°], a frontal-view face in a range of [−20°, 20°], and a right-view face in a range of [20°, 90°] are detected with respect to the Y-rotation.
  • When a face is rotated around the Z-axis, Z-rotation (left and right leaning) has a range of [−180°, 180°]. The Z-rotation is referred to as “in-plane rotation”.
  • With respect to Z-rotation, all rotation in the range of [−180°, 180°] are dealt. However, people can lean the face only in a range of [−45°, 45°] when standing. Accordingly, detection is performed with respect to rotation in the range of [−45°, 45°] in a basic mode and is performed with respect to rotation in the range of [−180°, 180°] in an extension mode.
  • In addition, with respect to the Z-rotation, a face may be defined to lean at intervals of 30° and 45°, which is respectively illustrated in FIGS. 4A and 4B. Hereinafter, a mode illustrated in FIG. 4A is referred to as a first Z-rotation mode and a mode illustrated in FIG. 4B is referred to as a second Z-rotation mode. In the Z-rotation, a left-leaned face, an upright face, and a right-leaned face are defined.
  • Table 1 shows the ranges of rotation angles of a face to be detected according to an embodiment of the present invention.
    TABLE 1
    Division
    X-rotation Y-rotation Z-rotation
    Description Up-and-down Left and right Left and right
    nodding rotation leaning
    Rotatable angle [−60°, 80°] [−180°, 180°] [−180°, 180°]
    Detec- Basic [−60°, 50°] [−90°, 90°] [−45°, 45°]
    tion mode
    target Extension [−60°, 50°] [−90°, 90°] [−180°, 180°]
    mode
  • Meanwhile, a face detection apparatus according to an embodiment of the present invention may detect a face using cascaded classifiers, each of which is trained with conventional appearance-based pattern recognition, i.e., an AdaBoost algorithm. The AdaBoost algorithm is an efficient learning algorithm that configures a plurality of simple and fast weak classifiers in a form of a weighted sum, thereby producing a single strong classifier which is fast and has a high success rate. Hereinafter, a strong classifier for detecting a particular face pose is referred to as a “face detector”.
  • The face detector discriminates a face from a non-face in an input image using a plurality of face patterns that it has learned. Accordingly, it is necessary to determine face patterns to be learned.
  • As described above, to detect a down-view face in the range of [−60°, −20°] and a frontal-view face in the range of [−20°, 50°] with respect to the X-rotation, two face detectors are needed.
  • In addition, to detect a left-view face in the range of [−90°, −20°], a frontal-view face in the range of [−20°, 20°], and a right-view face in the range of [20°, 90°] with respect to the Y-rotation, three face detectors are needed.
  • In the first Z-rotation mode, 12 face detectors are needed in the extension mode and three face detectors are needed in the basic mode. In the second Z-rotation mode, 8 face detectors are needed in the extension mode and two face detectors are needed in the basic mode.
  • Consequently, when all of the X-, Y-, and Z-rotations are considered in the first X-rotation mode, 2×3×3=18 face detectors are needed in the basic mode and 2×3×12=72 face detectors are needed in the extension mode.
  • When all of the X-, Y-, and Z-rotations are considered in the second X-rotation mode, 2×3×2=12 face detectors are needed in the basic mode and 2×3×8=48 face detectors are needed in the extension mode.
  • However, in the first and second Z-rotation modes, the number of face detectors for learning can be reduced by using rotation or mirroring (changing left and right coordinates), which is illustrated in FIG. 5.
  • For example, with respect to a frontal-view face in the first Z-rotation mode, when an upright face 502 is rotated by −90°, 90°, and 180°, face images 508, 520, and 514 are obtained. When a 30° left-leaned face 524 is rotated by −90°, 90°, and 180°, face images 506, 518, and 512 are obtained. In addition, when the 30° left-leaned face 524 is mirrored, a 30° right-leaned face 504 is obtained. When the 30° right-leaned face 504 is rotated by −90°, 90°, and 180°, face images 510, 522, and 516 are obtained. As a result, since faces other than the upright face 502 and the 30° left-leaned face 524 can be obtained through rotation or mirroring, 12 face detectors for a frontal-view face can be made by learning two face detectors.
  • In the same manner, as shown in FIG. 6, 12 face detectors can be made using three face detectors with respect to a left-view face. In addition, a right-view face can be obtained by mirroring the left-view face.
  • Consequently, when all of the X-, Y-, and Z-rotations are considered in the first X-rotation mode, 2 (a frontal-view and a down-view)x5=10 face detectors are needed to be learned in the basic and extension modes. Here, faces to be learned with respect to the frontal-view face are shown in FIG. 7.
  • Referring to FIG. 8, with respect to a frontal-view face in the second Z-rotation mode, when a right-leaned face 802 in the basic mode is rotated by −90°, 90°, and 180°, face images 806, 814, and 810 are obtained. When the right-leaned face 802 is mirrored, a left-leaned face 816 is obtained. When the left-leaned face 816 is rotated by −90°, 90°, and 180°, face images 804, 812, and 808 are obtained. Consequently, when only the right-leaned face 802 is learned, other faces can be obtained through rotation or mirroring. Accordingly, 8 face detectors for the frontal-view face can be made by learning only a single face detector.
  • In the same manner, referring to FIG. 9, 8 face detectors can be made using two face detectors with respect to a left-view face. In addition, a right-view face can be obtained by mirroring the left-view face.
  • Consequently, when all of the X-, Y-, and Z-rotations are considered in the second Z-rotation mode, 2 (a frontal-view and a down-view)x3=6 face detectors are needed to learn in the basic and extension modes. Here, faces to be learned with respect to the frontal-view face are shown in FIG. 10.
  • Table 2 shows the number of face detectors needed in an embodiment of the present invention.
    TABLE 2
    Number of Number of
    necessary face face detectors
    detectors to learn
    First mode Basic mode 18 10
    Extension mode 72 10
    Second mode Basic mode 12 6
    Extension mode 48 6
  • FIG. 11 is a block diagram of a face detection apparatus 1100 according to an embodiment of the present invention. The face detection apparatus 1100 includes an image sensing module 1120, a face detection module 1140, a first storage module 1160, and a second storage module 1180.
  • The image sensing module 1120 has an imaging function like a camera. The image sensing module 1120 senses an image of an object and provides the image to the face detection module 1140.
  • The first storage module 1160 stores images sensed by the image sensing module 1120 or images captured by a user and provides the stored images to the face detection module 1140 according to the user's request.
  • The face detection module 1140 detects a human face from an image received from the image sensing module 1120 or the first storage module 1160.
  • The second storage module 1180 stores an image of the detected human face. The image stored in the second storage module 1180 may be transmitted to a display apparatus 1182, a face recognition apparatus 1184, or other image processing apparatus through a wired/wireless network 1186.
  • The first storage module 1160 and the second storage module 1180 may be implemented as different storage areas in a physically single storage medium or may be implemented as different storage media, respectively.
  • In addition, the storage areas for the respective first and second storage modules 1160 and 1180 may be defined by a software program.
  • The term “module,” as used herein, means, but is not limited to, a software or hardware component, such as a Field Programmable Gate Array (FPGA) or Application Specific Integrated Circuit (ASIC), which performs certain tasks. A module may advantageously be configured to reside on the addressable storage medium and configured to execute on one or more processors. Thus, a module may include, by way of example, components, such as software components, object-oriented software components, class components and task components, processes, functions, attributes, procedures, subroutines, segments of program code, drivers, firmware, microcode, circuitry, data, databases, data structures, tables, arrays, and variables. The functionality provided for in the components and modules may be combined into fewer components and modules or further separated into additional components and modules.
  • FIG. 12 is a block diagram of an example of the face detection module 1140 illustrated in FIG. 11. The face detection module 1140 includes a controller 1142, a subwindow generator 1144, a first face searcher 1146, a second face searcher 1148, and a third face searcher 1150.
  • The subwindow generator 1144 generates a subwindow for an input image received from the image sensing module 1120 or the first storage module 1160. The subwindow is a portion clipped out of the input image in a predetermined size. For example, when the input image has a size of 320×240 pixels, if an image of 24×24 pixels is clipped, the clipped image will be a subwindow of the input image. Here, the subwindow generator 1144 defines a minimum subwindow size and increases the length or width of a subwindow step by step starting from the minimum subwindow size. In other words, the subwindow generator 1144 sequentially provides the first face searcher 1146 with subwindows generated while increasing the size of the subwindow step by step.
  • The first face searcher 1146, the second face searcher 1148, and the third face searcher 1150 perform operations to detect a face from each subwindow generated by the subwindow generator 1144.
  • The controller 1142 controls the operation of the subwindow generator 1144 according to whether a face is detected by the operations of the first through third face searchers 1146 through 1150.
  • Upon receiving a subwindow from the subwindow generator 1144, the first face searcher 1146 searches for a face in the subwindow using a predetermined algorithm. If a face is detected, the first face searcher 1146 transmits the subwindow to the second face searcher 1148. However, if no face is detected, the controller 1142 controls the subwindow generator 1144 to generate and transmit a new subwindow to the first face searcher 1146. The second face searcher 1148 searches for a face in the received subwindow using a predetermined algorithm. If a face is detected, the second face searcher 1148 transmits the subwindow to the third face searcher 1150. However, if no face is detected, the controller 1142 controls the subwindow generator 1144 to generate and transmit a new subwindow to the first face searcher 1146. The third face searcher 1150 searches for a face in the received subwindow using a predetermined algorithm. If a face is detected, the third face searcher 1150 stores the subwindow in a separate storage area (not shown). After face search is completely performed on all subwindows of the image provided by the image sensing module 1120 or the first storage module 1160, face detection information of the image is stored in the second storage module 1180 based on the stored subwindows. However, if no face is detected by the third face searcher 1150, the controller 1142 controls the subwindow generator 1144 to generate and transmit a new subwindow to the first face searcher 1146.
  • The algorithms respectively used by the first face searcher 1146, the second face searcher 1148, and the third face searcher 1150 to search for a face will be described with reference to FIGS. 13A through 13C.
  • FIG. 13A illustrates a conventional coarse-to-fine search algorithm, FIG. 13B illustrates a conventional simple-to-complex search algorithm, and FIG. 13C illustrates a parallel-to-separated search algorithm according to an embodiment of the present invention.
  • In the coarse-to-fine search algorithm, a whole-view classifier is made at an initial stage of a cascaded classifier and then classifiers for gradually narrower angles are made. When the coarse-to-fine search algorithm is used, a non-face is quickly removed in early stages so that entire detection time can be reduced. The whole-view classifier only searches for the shape of a face in a given subwindow using information that has been learned, regardless of the pose of the face.
  • In the simple-to-complex search algorithm, an easy and simple classifier is disposed at an earlier stage and a difficult and complex classifier is disposed at a latter stage to increase speed. Since most of non-faces are removed in an initial stage, a great effect can be achieved when the initial stage is made simple.
  • In the parallel-to-separated search algorithm according to an embodiment of the present invention, face detectors for all directions are arranged in parallel up to, for example, K-th stages, and face detectors for respective different directions are independently and separately arranged starting from a (K+1)-th stage. In the parallel arrangement, when face detection succeeds in one direction, a subsequent stage in the same direction is continued. However, when face detection fails in one direction, face detection is performed in a different direction. In the separated arrangement, when face detection in one direction succeeds, a subsequent stage in the same direction is continued. However, when face detection fails, a non-face is immediately determined and the face detection is terminated. When the parallel-to-separated search algorithm is used, the direction of a face in an input image is determined in an initial stage, and thereafter, a face or a non-face is determined only with respect to the direction. Accordingly, a face detector having high accuracy and fast speed can be implemented.
  • When the algorithms illustrated in FIGS. 13A through 13C are combined, a multi-view face detector illustrated in FIG. 14 can be made.
  • In FIG. 14, each block is a face detector detecting a face in a direction written in the block. An area denoted by “A” operates in the same manner as the left part, and thus a description thereof is omitted. A downward arrow indicates a flow of the operation when a face detector succeeds in detecting a face. A rightward arrow indicates a flow of the operation when a face detector fails in detecting a face.
  • For example, referring to FIGS. 12 and 14, upon receiving a subwindow from the subwindow generator 1144, the first face searcher 1146 discriminates a face from a non-face using a whole-view face detector based on already learned information in stage 1˜1′.
  • When a face is determined in stage 1˜1′, the first face searcher 1146 transmits the subwindow to the second face searcher 1148. The second face searcher 1148 performs stage 2˜2′ and stage 3˜4.
  • In stage 2˜2′, a frontal-view face and a down-view face are grouped with respect to the X-rotation and face detection is performed based on the already learned information. In stage 3˜4, an upright face, a left-leaned face, and a right-leaned face are grouped with respect to the Z-rotation and face detection is performed based on the already learned information.
  • Stage 1˜1′, stage 2˜2′, and stage 3˜4 are performed using the coarse-to-fine search algorithm. Face detectors performing stage 1˜1′, stage 2˜2′, and stage 3˜4 internally use the simple-to-complex search algorithm.
  • Up to stage M, faces in all directions are classified based on the already learned information. Here, up to stage K, when face detection succeeds, a subsequent downward stage is performed and when face detection fails, the operation shifts to a right face detector. After stage K, when face detection succeeds, a subsequent downward stage is performed, but when face detection fails, a non-face is determined and face detection on a current subwindow is terminated. Accordingly, only with respect to a subwindow reaching stage M, it is determined that a face is detected.
  • Stage 5˜K and stage K+1˜M are performed using the parallel-to-separated search algorithm. In addition, face detectors performing stage 5˜K and stage K+1˜M internally use the simple-to-complex search algorithm.
  • FIGS. 15A and 15B are flowcharts of a method of detecting a face according to an embodiment of the present invention. FIG. 16 is a different type of flowchart of the method according to the embodiment illustrated in FIGS. 15A and 15B.
  • The method of detecting a face according to an embodiment of the present invention will be described with reference to FIGS. 11, 12, and 15A through 16. Here, it is assumed that an image provided by the image sensing module 1120 or the first storage module 1160 is a frontal-view face defined with respect to the X-rotation. Accordingly, stage 2˜2′ shown in FIG. 14 is omitted, and a stage for detecting a whole-view face is referred to as stage 1˜2. In FIG. 6, W represents “whole-view”, U represents “upright”, L represents “30° left-leaned”, R represents “30° right-leaned”, “f” represents “fail” indicating that face detection has failed, “s” represents “succeed” indicating that face detection has succeeded, and NF represents “non-face”.
  • When the subwindow generator 1144 generates a subwindow in operation S1502, an initial value for detecting a face in the subwindow is set in operation S1504. The initial value includes parameters n, N1, N2, K, and M.
  • The parameter “n” indicates a stage in the face detection. The parameter N1 indicates a reference value for searching for a whole-view face. The parameter N2 indicates a reference value for searching for an upright-view face, a left leaned-view face, and a right leaned-view face defined with respect to the Z-rotation. The parameter M indicates a reference value for searching for a frontal-view face, a left-view face, and a right-view face defined with respect to the Y-rotation. The parameter K indicates a reference value for discriminating a stage for arranging face detectors separately from a stage for arranging face detectors in parallel in the parallel-to-separated search algorithm according to an embodiment of the present invention. Here, the initial value is set such that n=1, N1=2, N2=4, K=10, and M=25.
  • After the initial value is set, the first face searcher 1146 searches for a whole-view face in stage “n” in operation S1506, i.e., 1602. If a whole-view face is not detected, it is determined that no face exists in the subwindow. If a whole-view face is detected, the parameter “n” is increased by 1 in operation S1508. It is determined whether the value of “n” is greater than the value of N1 in operation S1510. If the value of “n” is not greater than the value of N1, the method goes back to operation S1506. Since the parameter N1 is set to 2 in the embodiment of the present invention, the first face searcher 1146 performs the simple-to-complex search algorithm on the whole-view face two times (16021604).
  • If it is determined that the value of “n” is greater than the value of N1 in operation S1510, the second face searcher 1148 searches for an upright-view face in stage “n” in operation S1512 (i.e., 1606). Here, the coarse-to-fine search algorithm is used.
  • If the upright-view face is not detected in operation S1512 (i.e., 1606), a left leaned-view face is searched for in the same stage “n” in operation S1560 (i.e., 1608). If the left leaned-view face is not detected in operation S1560, a right leaned-view face is searched for in the same stage “n” in operation S1570 (i.e., 1610). If the right leaned-view face is not detected in operation S1570, it is determined that no face exists in the current subwindow. If a face is detected in operation S1512 (1606), S1560 (1608), or S1570 (1610), the value of “n” is increased by 1 in operation S1514, S1562, or S1572, respectively, and it is determined whether the increased value of “n” is greater than the value of N2 in operation S1516, S1564, or S1574, respectively. If the value of “n” is not greater than the value of N2, the method goes back to operation S1512, S1560, or S1570. Since the value of N2 is set to 4 in the embodiment of the present invention, the second face searcher 1148 performs the simple-to-complex search algorithm on the upright-view face, the left leaned-view face, or the right leaned-view face two times (16061612, 16081614, or 16101616).
  • Hereinafter, for clarity of the description, it is assumed that a face is detected in operation S1512 and the value of “n” is greater than the value of N2 in operation S1516. Referring to FIG. 15B, the same operations as operations S1520 through S1554 are performed in an 1-block (S1566) and an 11-block (S1576).
  • The third face searcher 1150 searches for an upright frontal-view face in stage “n” in operation S1520 (i.e., 1618). If the upright frontal-view face is not detected in operation S1520 (1618), an upright left leaned-view face is detected in the same stage “n” in operation S1526 (i.e., 1620). If the upright left leaned-view face is not detected in operation S1526, an upright right leaned-view face is searched for in the same stage “n” in operation S1532 (i.e., 1622). If the upright right leaned-view face is not detected in operation S1532, face detection is continued in the 1-block (S1566) or the 11-block (S1576).
  • If a face is detected in operation S1520 (1618), S1526 (S1620), or S1532 (1622), the value of “n” is increased by 1 in operation S1522, S1528, or S1534, respectively, and it is determined whether the increased value of “n” is greater than the value of K in operation S1524, S1530, or S1535, respectively. If the value of “n” is not greater than the value of K, the method goes back to operation S1520, S1526, or S1532. Since the value of K is set to 10 in the embodiment of the present invention, the third face searcher 1150 performs the simple-to-complex search algorithm on the upright frontal-view face, the upright left leaned-view face, or the upright right leaned-view face up to a maximum of 6 times (16181624, 16201626, or 16221628).
  • Hereinafter, for clarity of the description, it is assumed that the upright frontal-view face is detected in operation S1520 and it is determined that the value of “n” is greater than the value of K in operation S1524.
  • The third face searcher 1150 searches for an upright frontal-view face in stage “n” in operation S1540 (i.e., 1630). If the upright frontal-view face is not detected in operation S1540 (1630), it is determined that no face exists in the current subwindow. If the upright frontal-view face is detected in operation S1540, the value of “n” is increased by 1 in operation S1542 and it is determined whether the increased value of “n” is greater than the value of M in operation S1544. If the increased value of “n” is not greater than the value of M, the method goes back to operation S1540. If the increased value of “n” is greater than the value of M, it is determined that a face exists in the current subwindow.
  • As described above, the third face searcher 1150 operates using the parallel-to-separated search algorithm according to an embodiment of the present invention and the conventional simple-to-complex search algorithm. In other words, face detectors for all directions are arranged in parallel up to stage K and are arranged separately from each other from stage K+1 to stage M, and the simple-to-complex search algorithm is used when a stage shifts.
  • Meanwhile, in an embodiment of the present invention, X-rotation faces are detected first, Z-rotation faces are detected next, and Y-rotation faces are detected finally. However, such order is just an example, and it will be obvious to those skilled in the art that the order may be changed in face detection.
  • According to the above-described embodiments of the present invention, any one of the X-rotation faces, Y-rotation faces and Z-rotation faces can be detected. In addition, since a pose estimator is not used prior to a multi-view face detector, an error of the pose estimator does not occur and accuracy and operating speed increase. As a result, an efficient operation can be performed.
  • The above-described embodiments of the present invention can be used for any field requiring face recognition such as credit cards, cash cards, digital social security cards, cards needing identity authentication, terminal access control, control systems in public places, digital album, and recognition of photographs of criminals. The present invention can also be used for security monitoring systems.
  • Although a few embodiments of the present invention have been shown and described, the present invention is not limited to the described embodiments. Instead, it would be appreciated by those skilled in the art that changes may be made to these embodiments without departing from the principles and spirit of the invention, the scope of which is defined by the claims and their equivalents.

Claims (22)

1. A method of detecting a multi-view face, comprising:
sequentially attempting to detect from an input image two mode faces among a first mode face made by up and down rotation of a face, a second mode face made by leaning a head to the left and right, and a third mode face made by left and right rotation of the face;
attempting to detect the remaining mode face that is not detected in the sequentially attempting; and
determining that a face is detected from the input image when the remaining mode face is detected in the attempting to detect the remaining mode,
wherein the attempting to detect the remaining mode comprises:
arranging face detectors for detectable directions in parallel and performing face detection; and
independently and separately arranging the face detectors for the detectable directions and performing face detection.
2. The method of claim 1, wherein the arranging face detectors comprises:
arranging the face detectors for the detectable directions in parallel;
performing face detection in the same direction using a more complex face detector when face detection succeeds in one direction; and
performing face detection in a different direction when face detection fails in one direction.
3. The method of claim 1, wherein the independently and separately arranging comprises:
independently and separately arranging the face detectors for the detectable directions;
performing face detection in the same direction using a more complex face detector when face detection succeeds in one direction; and
determining that a face is not detected from the input image when face detection fails.
4. The method of claim 1, wherein, in the sequentially attempting, sequential detection of the two mode faces is attempted using a coarse-to-fine search algorithm.
5. The method of claim 1, wherein in the sequentially attempting, detection of one of the two mode faces is attempted using a simple-to-complex search algorithm.
6. The method of claim 1, wherein the first mode face comprises a down-view face in a range of [−60°, −20°] around an X-axis and a frontal-view face in a range of [−20°, 50°] around the X-axis.
7. The method of claim 1, wherein the second mode face comprises an upright face and a leaned face made by rotating the upright face by −30° or 30° around a Z-axis in a range of [−45°, 45°] around the Z-axis and comprises the upright face, the leaned face, and other leaned faces obtained using the upright face and the leaned face in a range of [−180°, 180°] around the Z-axis.
8. The method of claim 1, wherein the second mode face comprises two leaned faces having a rotation angle of 45° therebetween in a range of [−45°, 45°] around a Z-axis and comprises the two leaned face and other leaned faces obtained using the two leaned faces in a range of [−180°, 180°] around the Z-axis.
9. The method of claim 1, wherein the third mode face comprises a left-view face in a range of [−90°, −20°] around a Y-axis, a frontal-view face in a range of [−20°, 20°] around the Y-axis, and a right-view face in a range of [20°, 90°] around the Y-axis.
10. An apparatus of detecting a multi-view face, including a face detection module, the module comprising:
a subwindow generator receiving an input image and generating a subwindow with respect to the input image;
a first face searcher receiving the subwindow and determining whether a whole-view face exists in the subwindow;
a second face searcher sequentially searching for two mode faces among a first mode face made by up and down rotation of a face, a second mode face made by leaning a head to the left and right, and a third mode face made by left and right rotation of the face when the first face searcher determines that the whole-view face exists in the subwindow;
a third face searcher searching for the remaining mode face that is not searched for by the second face searcher; and
a controller controlling the subwindow generator to generate a new subwindow when one of the first face searcher, the second face searcher, and the third face searcher does not detect a face.
11. The apparatus of claim 10, further comprising an image sensing module sensing an image of an object, wherein the input image received by the subwindow generator is the image received from the image sensing module.
12. The apparatus of claim 10, further comprising a storage module storing an image captured by a user, wherein the input image received by the subwindow generator is the image received from the storage module.
13. The apparatus of claim 10, further comprising a storage module storing a face image detected by the face detection module.
14. The apparatus of claim 10, wherein the third face searcher sequentially performs:
an operation of arranging face detectors for all directions in parallel when succeeding in face detection in one direction, performing face detection in the same direction using a more complex face detector, and performing face detection in a different direction when failing in face detection in one direction; and
an operation of independently and separately arranging the face detectors for all directions when succeeding in face detection in one direction, performing face detection in the same direction using a more complex face detector, and determining that a face is not detected from the input image when failing in face detection.
15. The apparatus of claim 10, wherein the first face searcher performs sequential detection of the two mode faces using a coarse-to-fine search algorithm.
16. The apparatus of claim 10, wherein the second face searcher performs detection of one of the two mode faces using a simple-to-complex search algorithm.
17. The apparatus of claim 10, wherein the first mode face comprises a down-view face in a range of [−60°, −20°] around an X-axis and a frontal-view face in a range of [−20°, 50°] around the X-axis.
18. The apparatus of claim 10, wherein the second mode face comprises an upright face and a leaned face made by rotating the upright face by −30° or 30° around a Z-axis in a range of [−45°, 45°] around the Z-axis and comprises the upright face, the leaned face, and other leaned faces obtained using the upright face and the leaned face in a range of [−180°, 180°] around the Z-axis.
19. The apparatus of claim 10, wherein the second mode face comprises two leaned faces having a rotation angle of 45° therebetween in a range of [−45°, 45°] around a Z-axis and comprises the two leaned face and other leaned faces obtained using the two leaned faces in a range of [−180°, 180°] around the Z-axis.
20. The apparatus of claim 10, wherein the third mode face comprises a left-view face in a range of [−90°, −20°] around a Y-axis, a frontal-view face in a range of [−20°, 20°] around the Y-axis, and a right-view face in a range of [20°, 90°] around the Y-axis.
21. The apparatus of claim 10, wherein the first, second, or third face searcher used uses cascaded classifiers, each of which is trained with an appearance-based pattern recognition algorithms.
22. A computer-readable storage medium encoded with processing instructions for causing a processor to execute a method of detecting a multi-view face, comprising:
sequentially attempting to detect from an input image two mode faces among a first mode face made by up and down rotation of a face, a second mode face made by leaning a head to the left and right, and a third mode face made by left and right rotation of the face;
attempting to detect the remaining mode face that is not detected in the sequentially attempting; and
determining that a face is detected from the input image when the remaining mode face is detected in the attempting to detect the remaining mode,
wherein the attempting to detect the remaining mode comprises:
arranging face detectors for detectable directions in parallel and performing face detection; and
independently and separately arranging the face detectors for the detectable directions and performing face detection.
US11/285,172 2004-12-07 2005-11-23 Method and apparatus for detecting multi-view faces Abandoned US20060120604A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020040102411A KR100643303B1 (en) 2004-12-07 2004-12-07 Method and apparatus for detecting multi-view face
KR10-2004-0102411 2004-12-07

Publications (1)

Publication Number Publication Date
US20060120604A1 true US20060120604A1 (en) 2006-06-08

Family

ID=36574269

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/285,172 Abandoned US20060120604A1 (en) 2004-12-07 2005-11-23 Method and apparatus for detecting multi-view faces

Country Status (2)

Country Link
US (1) US20060120604A1 (en)
KR (1) KR100643303B1 (en)

Cited By (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070122010A1 (en) * 2005-11-01 2007-05-31 Fujifilm Corporation Face detection method, apparatus, and program
US20070211918A1 (en) * 2006-03-10 2007-09-13 Fujifilm Corporation Target-image detecting apparatus and method of controlling same
US20080175447A1 (en) * 2007-01-24 2008-07-24 Samsung Electronics Co., Ltd. Face view determining apparatus and method, and face detection apparatus and method employing the same
EP1973064A2 (en) 2007-03-21 2008-09-24 Ricoh Company, Ltd. Object image detection method and object image detection device
US20080232698A1 (en) * 2007-03-21 2008-09-25 Ricoh Company, Ltd. Object image detection method and object image detection device
US20080232693A1 (en) * 2007-03-20 2008-09-25 Ricoh Company, Limited Image processing apparatus, image processing method, and computer program product
US20090322906A1 (en) * 2008-06-26 2009-12-31 Casio Computer Co., Ltd. Imaging apparatus, imaged picture recording method, and storage medium storing computer program
US20100166317A1 (en) * 2008-12-30 2010-07-01 Li Jiangwei Method, apparatus and computer program product for providing face pose estimation
US20100284622A1 (en) * 2007-11-23 2010-11-11 Samsung Electronics Co., Ltd. Method and apparatus for detecting objects
US20100329565A1 (en) * 2009-06-29 2010-12-30 Canon Kabushiki Kaisha Image processing apparatus image processing method, and control program to perform face-detection processing
JP2013074570A (en) * 2011-09-29 2013-04-22 Sanyo Electric Co Ltd Electronic camera
CN103577841A (en) * 2013-11-11 2014-02-12 浙江大学 Human body behavior identification method adopting non-supervision multiple-view feature selection
US20140133743A1 (en) * 2011-09-27 2014-05-15 Olaworks, Inc. Method, Apparatus and Computer Readable Recording Medium for Detecting a Location of a Face Feature Point Using an Adaboost Learning Algorithm
US20150043790A1 (en) * 2013-08-09 2015-02-12 Fuji Xerox Co., Ltd Image processing apparatus and non-transitory computer readable medium
US20150206311A1 (en) * 2014-01-20 2015-07-23 Apical Ltd. Method of object orientation detection
CN105488371A (en) * 2014-09-19 2016-04-13 中兴通讯股份有限公司 Face recognition method and device
US9697346B2 (en) * 2012-03-06 2017-07-04 Cisco Technology, Inc. Method and apparatus for identifying and associating devices using visual recognition
CN106921658A (en) * 2017-02-14 2017-07-04 上海斐讯数据通信技术有限公司 A kind of router device safety protecting method and system
US10025998B1 (en) * 2011-06-09 2018-07-17 Mobileye Vision Technologies Ltd. Object detection using candidate object alignment
US20190026538A1 (en) * 2017-07-21 2019-01-24 Altumview Systems Inc. Joint face-detection and head-pose-angle-estimation using small-scale convolutional neural network (cnn) modules for embedded systems
US10691925B2 (en) * 2017-10-28 2020-06-23 Altumview Systems Inc. Enhanced face-detection and face-tracking for resource-limited embedded vision systems

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100779171B1 (en) * 2006-08-18 2007-11-26 학교법인 포항공과대학교 Real time robust face detection apparatus and method

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6128397A (en) * 1997-11-21 2000-10-03 Justsystem Pittsburgh Research Center Method for finding all frontal faces in arbitrarily complex visual scenes
US20030108244A1 (en) * 2001-12-08 2003-06-12 Li Ziqing System and method for multi-view face detection
US20040186816A1 (en) * 2003-03-17 2004-09-23 Lienhart Rainer W. Detector tree of boosted classifiers for real-time object detection and tracking
US20050213810A1 (en) * 2004-03-29 2005-09-29 Kohtaro Sabe Information processing apparatus and method, recording medium, and program
US6959109B2 (en) * 2002-06-20 2005-10-25 Identix Incorporated System and method for pose-angle estimation
US20070086660A1 (en) * 2005-10-09 2007-04-19 Haizhou Ai Apparatus and method for detecting a particular subject

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH09171560A (en) * 1995-12-20 1997-06-30 Oki Electric Ind Co Ltd Device for detecting face inclination
KR20020022295A (en) * 2000-09-19 2002-03-27 장기화 Device And Method For Face Recognition Using 3 Dimensional Shape Information
US20030063781A1 (en) 2001-09-28 2003-04-03 Koninklijke Philips Electronics N.V. Face recognition from a temporal sequence of face images
JP2004005014A (en) 2002-05-30 2004-01-08 Symtron Technology Inc Face recognition system
JP2004062565A (en) 2002-07-30 2004-02-26 Canon Inc Image processor and image processing method, and program storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6128397A (en) * 1997-11-21 2000-10-03 Justsystem Pittsburgh Research Center Method for finding all frontal faces in arbitrarily complex visual scenes
US20030108244A1 (en) * 2001-12-08 2003-06-12 Li Ziqing System and method for multi-view face detection
US6959109B2 (en) * 2002-06-20 2005-10-25 Identix Incorporated System and method for pose-angle estimation
US20040186816A1 (en) * 2003-03-17 2004-09-23 Lienhart Rainer W. Detector tree of boosted classifiers for real-time object detection and tracking
US20050213810A1 (en) * 2004-03-29 2005-09-29 Kohtaro Sabe Information processing apparatus and method, recording medium, and program
US20070086660A1 (en) * 2005-10-09 2007-04-19 Haizhou Ai Apparatus and method for detecting a particular subject

Cited By (39)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7809173B2 (en) * 2005-11-01 2010-10-05 Fujifilm Corporation Face detection method, apparatus, and program
US20070122010A1 (en) * 2005-11-01 2007-05-31 Fujifilm Corporation Face detection method, apparatus, and program
US20070211918A1 (en) * 2006-03-10 2007-09-13 Fujifilm Corporation Target-image detecting apparatus and method of controlling same
US7873186B2 (en) * 2006-03-10 2011-01-18 Fujifilm Corporation Target-image detecting apparatus and method of controlling same
US20080175447A1 (en) * 2007-01-24 2008-07-24 Samsung Electronics Co., Ltd. Face view determining apparatus and method, and face detection apparatus and method employing the same
KR101330636B1 (en) * 2007-01-24 2013-11-18 삼성전자주식회사 Face view determining apparatus and method and face detection apparatus and method employing the same
US8363909B2 (en) 2007-03-20 2013-01-29 Ricoh Company, Limited Image processing apparatus, image processing method, and computer program product
US20080232693A1 (en) * 2007-03-20 2008-09-25 Ricoh Company, Limited Image processing apparatus, image processing method, and computer program product
US20080232698A1 (en) * 2007-03-21 2008-09-25 Ricoh Company, Ltd. Object image detection method and object image detection device
US8660317B2 (en) 2007-03-21 2014-02-25 Ricoh Company, Ltd. Object image detection method and object image detection device for detecting an object image from an input image
US20080253664A1 (en) * 2007-03-21 2008-10-16 Ricoh Company, Ltd. Object image detection method and object image detection device
EP1973064A2 (en) 2007-03-21 2008-09-24 Ricoh Company, Ltd. Object image detection method and object image detection device
US8254643B2 (en) 2007-03-21 2012-08-28 Ricoh Company, Ltd. Image processing method and device for object recognition
US8666175B2 (en) 2007-11-23 2014-03-04 Samsung Electronics Co., Ltd. Method and apparatus for detecting objects
US20100284622A1 (en) * 2007-11-23 2010-11-11 Samsung Electronics Co., Ltd. Method and apparatus for detecting objects
US20090322906A1 (en) * 2008-06-26 2009-12-31 Casio Computer Co., Ltd. Imaging apparatus, imaged picture recording method, and storage medium storing computer program
US8004573B2 (en) * 2008-06-26 2011-08-23 Casio Computer Co., Ltd. Imaging apparatus, imaged picture recording method, and storage medium storing computer program
US8396263B2 (en) * 2008-12-30 2013-03-12 Nokia Corporation Method, apparatus and computer program product for providing face pose estimation
CN102272774A (en) * 2008-12-30 2011-12-07 诺基亚公司 Method, apparatus and computer program product for providing face pose estimation
WO2010076621A1 (en) * 2008-12-30 2010-07-08 Nokia Corporation Method, apparatus and computer program product for providing face pose estimation
US20100166317A1 (en) * 2008-12-30 2010-07-01 Li Jiangwei Method, apparatus and computer program product for providing face pose estimation
US8849035B2 (en) * 2009-06-29 2014-09-30 Canon Kabushiki Kaisha Image processing apparatus image processing method, and control program to perform face-detection processing
US20100329565A1 (en) * 2009-06-29 2010-12-30 Canon Kabushiki Kaisha Image processing apparatus image processing method, and control program to perform face-detection processing
US10025998B1 (en) * 2011-06-09 2018-07-17 Mobileye Vision Technologies Ltd. Object detection using candidate object alignment
US9202109B2 (en) * 2011-09-27 2015-12-01 Intel Corporation Method, apparatus and computer readable recording medium for detecting a location of a face feature point using an Adaboost learning algorithm
US9563821B2 (en) 2011-09-27 2017-02-07 Intel Corporation Method, apparatus and computer readable recording medium for detecting a location of a face feature point using an Adaboost learning algorithm
EP2763078A4 (en) * 2011-09-27 2016-04-13 Intel Corp Method, apparatus and computer readable recording medium for detecting a location of a face feature point using an adaboost learning algorithm
US20140133743A1 (en) * 2011-09-27 2014-05-15 Olaworks, Inc. Method, Apparatus and Computer Readable Recording Medium for Detecting a Location of a Face Feature Point Using an Adaboost Learning Algorithm
JP2013074570A (en) * 2011-09-29 2013-04-22 Sanyo Electric Co Ltd Electronic camera
US9697346B2 (en) * 2012-03-06 2017-07-04 Cisco Technology, Inc. Method and apparatus for identifying and associating devices using visual recognition
US20150043790A1 (en) * 2013-08-09 2015-02-12 Fuji Xerox Co., Ltd Image processing apparatus and non-transitory computer readable medium
CN103577841A (en) * 2013-11-11 2014-02-12 浙江大学 Human body behavior identification method adopting non-supervision multiple-view feature selection
US9483827B2 (en) * 2014-01-20 2016-11-01 Apical Ltd. Method of object orientation detection
US20150206311A1 (en) * 2014-01-20 2015-07-23 Apical Ltd. Method of object orientation detection
CN105488371A (en) * 2014-09-19 2016-04-13 中兴通讯股份有限公司 Face recognition method and device
CN106921658A (en) * 2017-02-14 2017-07-04 上海斐讯数据通信技术有限公司 A kind of router device safety protecting method and system
US20190026538A1 (en) * 2017-07-21 2019-01-24 Altumview Systems Inc. Joint face-detection and head-pose-angle-estimation using small-scale convolutional neural network (cnn) modules for embedded systems
US10467458B2 (en) * 2017-07-21 2019-11-05 Altumview Systems Inc. Joint face-detection and head-pose-angle-estimation using small-scale convolutional neural network (CNN) modules for embedded systems
US10691925B2 (en) * 2017-10-28 2020-06-23 Altumview Systems Inc. Enhanced face-detection and face-tracking for resource-limited embedded vision systems

Also Published As

Publication number Publication date
KR20060063286A (en) 2006-06-12
KR100643303B1 (en) 2006-11-10

Similar Documents

Publication Publication Date Title
US20060120604A1 (en) Method and apparatus for detecting multi-view faces
Ballard Animat vision
Kollreider et al. Real-time face detection and motion analysis with application in “liveness” assessment
US7885531B2 (en) Apparatus, medium, and method for photographing based on face detection
Hadid et al. Face and eye detection for person authentication in mobile phones
EP2336949B1 (en) Apparatus and method for registering plurality of facial images for face recognition
Wu et al. Simultaneous face detection and pose estimation using convolutional neural network cascade
Bileschi et al. Advances in component based face detection
EP1716519A2 (en) Face recognition system
TWI745818B (en) Method and electronic equipment for visual positioning and computer readable storage medium thereof
US20140341430A1 (en) Method and Device for Detecting Face, and Non-Transitory Computer-Readable Recording Medium for Executing the Method
WO2023169282A1 (en) Method and apparatus for determining interaction gesture, and electronic device
US20050276469A1 (en) Method for detecting face region using neural network
Mahbub et al. Pooling facial segments to face: The shallow and deep ends
Liao et al. Uniface: a unified network for face detection and recognition
Yüksel et al. Convolutional and recurrent neural networks for face image analysis
Srivastava et al. Face Verification System with Liveness Detection
Sultani et al. Action recognition in real-world videos
Sandoval et al. On the Use of a Low-Cost Embedded System for Face Detection and Recognition
Ren et al. Real-time head pose estimation on mobile platforms
Yu et al. SmartPartNet: Part-Informed Person Detection for Body-Worn Smartphones
Chai et al. Learning flexible block based local binary patterns for unconstrained face detection
Li et al. Activity Recognition
Hogue et al. Active Stereo Vision
Goldmann et al. Robust face detection based on components and their topology

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KIM, JUNG-BAE;PARK, CHAN-MIN;REEL/FRAME:017276/0745

Effective date: 20051118

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION