APPARATUS AND METHOD FOR GENERATING SYNTHETIC FACE AGE BASED ON SHAPE
INFORMATION ABOUT FACE IMAGE
Technical Field The present invention relates to an apparatus and method for generating a synthetic face image, and more particularly, to an apparatus and method for generating a synthetic face image based on shape information about an input face image.
A face image has generally been used as a medium of showing personal character most fully and allowing easy and natural communication. Such a face image is used in applications such as exit and entry control/security systems, criminal searching/montage composing systems, computer interfacing, animations, and games. Representative applications using technology of synthesizing face images are generation of a character image and makeup design.
A caricature of a face image, which is a sort of a character image, is produced based on the features of a person's face. Accordingly, a caricature of a face image is not only used for producing animations or entertaining programs but also used as a symbol or icon representing a person, as a unique signature in PC communications or E-mail, and as an avatar of a user in the virtual reality.
Background Art
In order to generate caricatures, conventionally, professional painters manually draw caricatures, or face images are automatically processed using digital filters. Here, a method of processing an image using digital filters gives an input image an effect of a caricature manually created as a whole by realizing effects of watercolor painting or charcoal drawing using combinations of filters having appropriate effects on an input image.
When a professional painter draws a caricature manually, a natural and highly perfect caricature can be obtained, but it takes a long time and it is difficult to maintain consistent quality. Accordingly, this method is restricted to particular conditions. An image processing technique using digital filters can be used for images picked up in an environment in which lighting and backgrounds are limited, so for simple two-dimensional images in which an object is not discriminated from a background, the quality of output images is depending on lighting or other changes in an environment. No proper methods for supplementing this technique have been proposed. Moreover, since shape information about an object is not separately generated in a conventional caricature generation method, it is very complex to perform a revising operation such as exaggerating of the features of the face in a generated caricature or changing of a facial expression, and it is nearly impossible to restore a face image or extend the generated caricature to a three-dimensional avatar.
In the case of makeup design, conventionally, a consumer determines a style indirectly based on pictures of models having on makeup. Recently, a makeup designing method using a computer has been introduced. This method allows to a user to apply products to sample model images in various ways but does not provide natural effects that can be obtained when directly applying makeup to the face image of a consumer. In other words, even if the same makeup product is used, color can be different depending on various and complex conditions such as shades and reflected light due to illumination and features of the shape of the face, so it is nearly impossible to naturally derive the makeup effects on the face image of a consumer from the makeup effects on a sample model image.
Disclosure of the Invention
To solve the above problems, it is an object of the present invention to provide an apparatus and method for generating a synthetic face image based on shape information about a face image, in which shape information about a face is extracted from an input face image, and a synthetic face image is generated based on the extracted shape information, so that a user can obtain a natural and elaborate caricature image, view a makeup design image using his/her face image in advance, easily add a variety of accessories to the synthetic image, easily change the synthetic image, and view the result image of adding accessories to or changing the synthetic image in real time.
To achieve the object of the invention, there is provided an apparatus ■ for synthesizing a new face image based on shape information about an input face image. The apparatus includes a user interfacing device for receiving face image information and a user control command, transmitting the same to an image processing device, receiving synthesized face image information from the image processing device, and outputting or storing the synthesized face image information according to the user control command; and an image processing device for extracting shape information, which is represented by a deformation field with respect to a predetermined reference image, and texture information, which is color or brightness information about an input image mapped on the reference image, about an input face image from face image information transmitted from the user interfacing device, and for transforming a texture image selected from among texture images, which are stored in an image database in advance and have the same shape as that of the reference image, or an image generated by performing a weighted summation of the selected texture image and a texture image reflecting the extracted texture information, using the shape information about the input face image according to the user control command, thereby generating a synthetic face image.
There is also provided a method for synthesizing a new face image based on shape information about an input face image. The method includes the steps of (a) extracting shape information, which is represented by a deformation field with respect to shape information about a predetermined reference image, and texture information, which is color or brightness information about an input image mapped on the reference image about an input face image from input face image information; and (b) transforming a texture image selected from among texture images, which are stored in an image database in advance and have the same shape as that of the reference image, or an image generated by performing a weighted summation of the selected texture image and a texture image reflecting the extracted texture information, using the shape information about the input face image according to a user control command, thereby generating a synthetic face image.
Brief Description of the Drawings
FIG. 1 A is a block diagram of the functional configuration of a first embodiment of a synthetic face image generating apparatus according to the present invention; FIG. 1 B is a block diagram of the functional configuration of a second embodiment of a synthetic face image generating apparatus according to the present invention;
FIG. 2 is a block diagram of the mechanical configuration of a computer system in which the first and second embodiments of the present invention are implemented;
FIG. 3 is a basic flowchart of a procedure of generating a synthetic face image according to the present invention;
FIG. 4 is a detailed flowchart of the step of extracting face information shown in FIG. 3; FIG. 5 is a flowchart of a procedure of generating a caricature
image according to a face image synthesizing method performed by an apparatus for generating a synthetic face image according to the present invention;
FIG. 6 is a flowchart of a procedure of generating a caricature image according to a sample image replacing method performed by an apparatus for generating a synthetic face image according to the present invention; and
FIG. 7 is a flowchart of a makeup design procedure performed by an apparatus for generating a synthetic face image according to the present invention.
Best mode for carrying out the Invention
Hereinafter, embodiments of the present invention will be described in detail with reference to the attached drawings. FIGS. 1A and 1B are block diagrams of a first embodiment 1 and a second embodiment 40, respectively, of a synthetic face image generating apparatus based on shape information about a face image according to the present invention. The first embodiment 1 shown in FIG. 1 includes one or more user interfacing devices 10a and 10b, a communication network 20, and an image processing device 30, thereby operating in a network environment. The second embodiment 40 shown in FIG. 2 includes a user interfacing device 50 and an image processing device 60, thereby operating in a single computer system.
The user interfacing devices 10a and 10b and the image processing device 30 of the first embodiment 1 , and the second embodiment 40 are realized as a computer system 70 including a computer 72 with at least one central processing unit (CPU) 74 and a memory unit 73, an input device 75, and an output device 76, as shown in FIG. 2. The elements of the computer system 70 are connected to one another through one or more bus structures 77.
The CPU 74 includes an arithmetical and logical unit (ALU) 741 for performing an arithmetical operation and a logical operation, a register set 742 for temporarily storing data and commands, and a control unit 743 for controlling the operations of the computer system 70. The CPU 74 used in the present invention is not restricted to a particular structure manufactured by a particular company but can be any type of processor having the above basic structure.
The memory unit 73 includes a high-speed main memory 731 and an auxiliary memory 732 for storing data for a long time. The main memory 731 may include a random access memory (RAM) semiconductor chip and a read only memory (ROM) semiconductor chip. The auxiliary memory 732 may include a floppy disc, a hard disc, a CD-ROM, a flash memory, and a device for storing data using electricity, magnetism, light or other recording medium. The main memory 731 may include a video display memory for displaying an image through a display device. It will be understood by those skilled in the field of the present invention that the memory unit 73 can include a variety of replaceable elements that have different storing performance.
The input device 75 may include a keyboard, a mouse, and a physical converter (for example, a microphone). The output device 76 may include a display unit, a printer, and a physical converter (for example, a speaker). Alternatively, a network interface, a modem, or the like can be used as an input/output device.
The computer system 70 is provided with an operating system and at least one application program. An operating system is a series of software controlling the operations of the computer system 70 and assignment of resources. An application program is a series of software performing work, which is requested by a user, using available computer resources through the operating system. This software is stored in the memory unit 73. Consequently, a computer-basis
synthetic face image generating apparatus according to the present invention is realized as the combination of the computer system 70 and one or more application programs which are installed and operate in the computer system 70. The first embodiment 1 shown in FIG. 1A has the same functions as the second embodiment 40 shown in FIG. 1 B, with the exception that the embodiment 1 further includes communication processors 14 and 31 for data transmission through the communication network 20. Thus, the present invention will be described on the basis of the first embodiment 1.
Referring to FIG. 1A, each of the user interfacing devices 10a and 10b receives face image information and a user control command from a user, receives a synthesized image in response to the user control command, and revises, stores or outputs the synthesized image. Each of the user interfacing devices 10a and 10b includes an image information input unit 11 , a user command input unit 12, an input/output controller 13, a communication processor 14, an image revision unit 15, an image storage unit 16, and an output unit 17.
The image information input unit 11 receives face image information from a user and can be realized as, for example, a scanner or a digital camera. In addition, the image information input unit 11 may include a plurality of cameras for receiving images picked up at different angles and a camera supplementary device such as a light adjustor. Since the image information input unit 11 as an element of the present invention should be considered in the functional aspect, the image information input unit 11 should be largely construed as including not only the input device 75 of FIG. 2 but also the auxiliary memory 732 of FIG. 2 which stores face image information in advance.
The user command input unit 12 receives user control commands (for example, user information, a face image synthesis control signal,
and an image revision control signal) from a user. The user command input unit 12 can be realized as a device such as a keyboard, a mouse or a touch screen through which a user can input a command or information. The input/output controller 13 controls face image information input through the image information input unit 11 and a user control command input through the user command input unit 12 to be transmitted to the image processing device 30 through the communication processor 14. The input/output controller 13 also receives new image information which is synthesized by the image processing device 30 in response to the user control command and controls the new image information to be revised, stored or output.
The communication processor 14 is connected to the input/output controller 13 and transmits data to or receives data from the image processing device 30 through the communication network 20. For example, the communication processor 14 can be realized as a device including a serial/parallel port, a universal serial bus (USB) port or an IEEE1394 port for internal connection, and an Ethernet card for transmitting and receiving data including image information through Internet.
The image revision unit 15 is connected to the input/output controller 13 and revises the angle, size or texture of an image corresponding to image information synthesized by and transmitted from the image processing device 30 in response to a user control command input through the user command input unit 12.
The image storage unit 16 corresponds to the auxiliary memory 732 of FIG. 2 and stores new image information synthesized by and transmitted from the image processing device 30 or image information revised by the image revision unit 15 under the control of the input/output controller 13.
The output unit 17 corresponds to the output device 76 of FIG. 2. Under the control of the input/output controller, the output unit 17 displays user interface display information used for receiving a user control command required when the image processing device 30 synthesizes a new image and displays or prints new image information synthesized by and transmitted from the image processing device 30 or image information revised by the image revision unit 15.
The communication network 20 transmits data between one or more user interfacing devices 10a and 10b and the image processing device 30 in the first embodiment 1 of the present invention shown in FIG. 1A. The communication network 20 can be realized as one of a variety of types of networks such as wire/wireless Internet, local area networks and private lines.
In the first embodiment 1 of the present invention shown in FIG. 1 , the image processing device 30 processes image information transmitted from one or more user interfacing devices 10a and 10b, and synthesizes a new image based on the image information in response to a user control command, and transmits the synthesized image to a corresponding user interfacing device. The image processing device 30 includes a communication processor 31 , an image processor 32, and an image database (DB) 33.
The communication processor 31 transmits data to and receives data from one or more user interfacing devices 10a and 10b. Corresponding to the communication processor 14 in each of the user interfacing devices 10a and 10b, the communication processor 31 can be realized as a device including an Ethernet card transmitting and receiving data including image information through Internet and a serial/parallel port, a universal serial bus (USB) port or an IEEE1394 port for internal connection. The image processor 32 extracts shape information about an
input face, image, indicated as a deformation field with respect to a reference image, and texture information, i.e., color or brightness information, about the input face image mapped on the reference image from face image information transmitted from the user interfacing devices 10a and 10b, analyzes a user control command transmitted from the user interfacing devices 10a and 10b to interpret a user's request, and synthesizes a new face image using the extracted shape information, the extracted texture information and a variety of images stored in the image DB 33 according to the interpreted user's request. The image processor 32 includes a face information extractor 321 , a face image synthesizer 322, a partial image replacing unit 323, and an accessory image adding unit 324.
The face information extractor 321 extracts from face image information transmitted from the user interfacing devices 10a and 10b shape information about an input face image, indicated as a deformation field with respect to a reference image, and texture information, i.e., color or brightness information, of the input face image mapped on the reference image using the shape information.
The face image synthesizer 322 synthesizes a new face image by transforming a texture image selected from among texture images stored in the image DB 33 in response to a user control command or transforming an image, which is generated by performing a weighted summation of the selected texture image and a texture image reflecting texture information extracted by the face information extractor 321 , using shape information about an input image, which is extracted by the face information extractor 321.
The partial image replacing unit 323 replaces a portion or the entire area of a new face image synthesized by the face image synthesizer 322 with a sample image having the highest similarity among sample images stored in the image DB 33.
The accessory image adding unit 324 adds an accessory image, which is selected from among accessory images stored in the image DB 33 according to a user control command, to a face image synthesized by the face image synthesizer 322. The image DB 33 previously stores image information necessary for processing an input face image in the image processor 32. The image DB 33 includes a face model DB 331 , an additional image DB 332, a sample image DB 333, a makeup image DB 334, and an accessory image DB 335. The face model DB 331 stores various types of information (a shape average, a texture average, shape eigenvectors, and texture eigenvectors which are previously obtained with respect to a plurality of model face images) used for extracting shape information and texture information, based on a reference image, from an input face image in the face information extractor 321. The various types of information stored in the face model DB 331 will be described in detail later when FIG. 4 is described.
The additional image DB 332 stores information about caricature images which have the same shape as that of a reference image but have different styles of texture information such as an animation style, a sketch style and a watercolor printing style.
The sample image DB 333 stores information about various caricature sample images including changes in a shape and a facial expression at each particular portion of a face image. The makeup image DB 334 stores information about makeup design images which have the same shape as that of a reference image and are expressed with texture information expressing a variety of sample makeups.
The accessory image DB 335 stores information about images such as glasses, hairstyles, hats, earrings, and bodies which can be
added to a synthesized face image.
In the first embodiment 1 of the present invention shown in FIG. 1A, one or more user interfacing devices 10a and 10b are connected to a single image processing device 30 through the communication processors 14 and 31 and the communication network 20. Alternatively, as in the second embodiment 40 of the present invention shown in FIG. 1 B, the user interfacing device 50 and the image processing device 60 can be integrated and operated within a single computer system 70.
The basic operations of the synthetic face image generating apparatuses 1 and 40 based on shape information about a face image according to the present invention will be described with reference to FIG. 3.
The face information extractor 321 or 621 of the image processor 32 or 62 receives an input face image from the user interfacing device 10a, 10b or 50 in step S10, and extracts shape information about the input face image with respect to a predetermined reference image and texture information, i.e., color or brightness information, of the input face image mapped on the reference image in step S11.
In step S12, the face image synthesizer 322 or 622 of the image processor 32 or 62 synthesizes a new face image using the texture information based on the reference image and the shape information about the input face image extracted by the face information extractor 321 or 621 , in response to a user control command (a face image synthesis control signal) received from the user interfacing device 10a, 10b or 50. In other words, the face image synthesizer 322 or 622 restores the shape of the input face image using the shape information extracted from the input face image and warps the restored shape of the input face image using the extracted texture information, thereby synthesizing a user's face image. Here, by appropriately changing or replacing the texture information based on the reference image, which is
used for synthesizing a face image, different new images having the shape of the input face image can be synthesized.
The synthesized face image is transmitted to the user interfacing device 10a, 10b or 50 and displayed through the output unit 17 or 57. The user command input unit 12 or 52 of the user interfacing device 10a, 10b or 50 receives a user control command indicating whether to change the shape information about the displayed face image from the user in step S13.
When the user control command instructs to change the shape information, the shape information about the input face image is changed according to the user control command (for example, a control signal instructing partial transform such as enlarging or reducing a particular portion of the displayed face image by dragging the portion with a mouse or a control signal instructing entire transform such as exaggerating the entire face using a slide bar), and the operation goes back to step S12 to synthesize a new face image.
When the user does not want to change the shape information in step S13, in response to an additional user control command, the accessory image adding unit 324 or 624 adds various accessory images stored in the image DB 33 or 63 to the face image synthesized in step S12, or the partial image replacing unit 323 or 623 replaces a particular portion of the face image synthesized in step S12 with various sample images stored in the image DB 33 or 63, thereby adding various additional effects to the synthesized face image in step S14. Thereafter, the face image synthesized by the image processing device 30 or 60 is transmitted to the user interfacing device 10a, 10b or 50 and displayed for the user. The image revision unit 15 or 55 finally revises the synthesized face image in response to a user control command (an image revision control signal) received through the user command input unit 12 or 52 in step S15. The synthesized face image
revised by the image revision unit 15 or 55 is stored in the image storage unit 16 or 56, or displayed or printed by the output unit 17 or 57 in step S16.
The face information extracting step S11 of FIG. 3 can be summarized as a procedure of obtaining shape information S,„ and texture information T,„ from an input face image based on a face model.
In the present invention, shape information about a face image is expressed by a deformation field with respect to a reference image, and texture information about a face image is expressed by color or brightness information about an input image mapped on the reference image. In other words, shape information S about a face image is defined as the positional difference between each point/?, on a reference image (z'=1 , ..., n; n is the number of predetermined points on the reference image) and each corresponding point on the face image. Texture information T about a face image is defined as the color or brightness value of each point on an input image with respect to each corresponding point p, (t-1, ..., ή) on a reference image. While a synthetic image obtained using a shape average and a texture average is used as a reference image in an embodiment of the present invention, a reference image which can be used in the present invention is not restricted to the above embodiment. Any one among m face images prepared previously can be used as a reference image.
Face models stored in the face model DB 331 or 631 are previously obtained according to the following procedure. First, shape information S, (/
'=1 , ..., m) and texture information 2} ( =1 , ..., m) are extracted from the previously prepared m face images based on a reference image. Thereafter, a shape average S composed of averages of shape information S
7 (/=1 , ..., m) at each point pi (t=1 , ..., ) and a texture average f composed of averages of texture information T
j
, ..., ) are obtained. Then a
covariance C
s of a shape difference S
J = S
J -S (/
'=1, ..., m) and a covariance C
τ of a texture difference T
J = T
J -T (/-1 , ..., m) are obtained.
These obtained values are processed by principal component analysis, thereby obtaining the shape eigenvector s, (z=1 , ..., mΛ) and texture eigenvector t, (z'=1 , ..., -1) of a covariance of the m face models. Therefore, a face image can be expressed by Equation (1) based on the shape eigenvector s, (/=1, ..., mΛ) and the texture eigenvector t, (z-1, ..., m- ).
S = S + ∑als„ T = T + ∑βfl ...(1)
(=1 ;=1
(Here, ά, β e Rm~l , and m is the number of face models) Through such a procedure, the shape average S , the texture average f , the shape eigenvector s, (z-1 , ..., m-1), and texture eigenvector t, (z-1 , ..., m-\) are stored in the face model DB 331 or 631 and used for extracting shape information and texture information about an input face image.
The face information extracting step S11 of FIG. 3 will be described in detail with reference to FIG. 4. In step S111 of FIG. 4, the input face image is normalized. More specifically, predetermined feature points (for example, a central point of each pupil and a central point of a lip) are extracted from an input face image, and then the input face image is moved up, down, to the left and to the right and the size thereof is adjusted so that the positions of the extracted feature points of the input face image can be located at the corresponding feature points of a reference image. Such an image normalizing step can be automatically performed through predetermined software or can be manually performed in response to a control command input by the user.
Since a detailed procedure thereof is beyond the scope of the present invention, a description thereof will be omitted.
In step S112, shape information is estimated. More specifically, a hierarchical, gradient-based optical flow algorithm (Lucas and Kanade) is applied to the normalized input face image obtained in step S111 and to the reference image (or a synthetic texture estimation image Tιn_k having the same shape as that of the reference image), thereby estimating shape information Sm_prej (the value of the positional difference between points on the normalized input face image and corresponding points on the reference image) based on the reference image. By using the hierarchical, gradient-based optical flow algorithm, the correspondence between two similar images can be expressed by an optical flow using the intensity of the two images.
The shape information Sιn_pre j obtained according to a hierarchical, gradient-based optical flow algorithm in step S112 may include an error value due to a light or shadow in the input face image. Accordingly, in step S1 13, compensation is performed on the shape information Sm_prej.
More specifically, the error value is compensated for by sequentially performing linear decomposition based on the shape eigenvector s, (z-1 , ..., mΛ) and linear superposition with respect to the shape information Smj}rej, thereby obtaining the error compensated shape information Sm correCtj. Here, to increase the degree of freedom in transform, it is preferable to use as the result value of the shape information a weighted sum Sm_ι of the estimated shape information Sm_preJ obtained in step S112 and the compensated shape information Sm_co recij obtained in step S113, as calculated by Equation (2).
(Here, Q≤ W≤ X) In step S1 14, backward warping is performed. More specifically,
the input face image is transformed into the reference image using the shape information Smj obtained in step S113. This procedure is referred to as "backward warping".
In step S115, linear decomposition is performed on texture information about the backward warped image based on the texture eigenvector t, (t-1, ..., mΛ), and then linear superposition is performed on the result of linear decomposition, thereby obtaining texture information T,„j about the input face image.
Thereafter, the normalized input face image obtained in step S112 is replaced with the input face image transformed into the reference image in step S114, and the reference image is replaced with a texture image having the same shape as that of the reference image. Then, steps S112 through S116 are repeated, thereby obtaining S,„_k (/=1 , ...). In other words, in a k-\h repetition, S,„j, Smj)rej, and S,„_co,τβc/ _/ in steps S112 and S113 are replaced with SιnJ, Smj)n , and Sm_correct , respectively, and S,„ i in step S114 is replaced with Sm(=S,„j+...+ Sιn ). In addition, Tmj in step S115 is replaced with T,„j. TmJ obtained in the last repetition becomes the final texture information T,„ about the input face image. Such a repetition is performed until the vector norm |SOT _k | of S,„ is less than a predetermined threshold value or until the number of repetitions is equal to or greater than a predetermined number, thereby obtaining shape information S,„ about the input face image based on the reference image in step S117.
Thereafter, the input face image can be restored using the shape information S,„ and the texture information Tm about the input face image, which is obtained based on the reference image. In other words, the texture information T,„ about the input face image obtained based on the reference image is transformed according to the shape information S,„ about the input face image obtained based on the reference image, thereby synthesizing the input face image.
Representative examples in which the characteristics of such a synthetic image can be utilized are generation of a caricature of a face image and makeup design. A method of generating a caricature of a face image is divided into a face image synthesizing method and a sample image replacement method.
A procedure of generating a caricature image using the face image synthesizing method performed in the synthetic face image generating apparatuses 1 and 40 based on the shape information about a face image according to the present invention will be described with reference to FIG. 5.
The face information extractor 321 or 621 of the image processor 32 or 62 receives an input face image from the user interfacing device 10a, 10b or 50 in step S20, and extracts shape information Sin and texture information Tin about the input face image based on a predetermined reference image in step S21.
Thereafter, the face image synthesizer 322 or 662 presents different styles of caricature images (for example, an animation style, a sketch style, and a watercolor printing style) stored in the additional image DB 332 or 632 to a user through the user interfacing device 10a, 10b or 50 so that the user can select a desired style of caricature in step S22. Here, different styles of caricature images stored in the additional image DB 332 or 632 have the same shape as that of the reference image.
In step S23, the face image synthesizer 322 or 622 synthesizes the selected style of caricature image or an image, which is generated by performing a weighted summation of the selected caricature image and an image reflecting the texture information Tιn about the input face image, with the shape information Sin about the input face image, thereby generating a caricature image reflecting the shape information about the user's face.
In step S24, the synthesized caricature image is transmitted to the user interfacing device 10a, 10b or 50 and displayed through the output unit 17 or 57, and the user command input unit 12 or 52 of the user interfacing device 10a, 10b or 50 receives a user control command indicating whether to change the shape information about the displayed caricature image from the user.
When the user control command instructs to change the shape information, the shape information Sin is changed according to the user control command (for example, a control signal instructing partial transform such as enlarging or reducing a particular portion of the displayed face image by dragging the portion with a mouse or a control signal instructing entire transform such as exaggerating the entire face using a slide bar), and the operation goes back to step S22 to synthesize a new caricature image. In step S25, in response to an additional user control command, the accessory image adding unit 324 or 624 retrieves various accessory images (glasses, hairstyles, hats, earrings, and bodies) stored in the accessory image DB 335 or 635, and adds the same to the caricature image. When the accessory image adding unit 324 or 624 adds an accessory image, more natural result can be obtained by automatically performing a size and position adjustment using the shape information Sin extracted in step S21. In addition, the partial image replacement unit 323 or 623 replaces a particular portion of the caricature image with a sample image retrieved from the sample image DB 333 or 633, thereby expressing happiness, sadness or anger on the caricature image. Moreover, a moving picture effect can be accomplished by using animation frames expressing changes in the face.
Thereafter, the caricature image synthesized by the image processing device 30 or 60 is transmitted to the user interfacing device 10a, 10b or 50 and displayed for the user. The image revision unit 15
or 55 finally revises the synthesized caricature image in response to a user control command (an image revision control signal) received through the user command input unit 12 or 52 in step S15. The synthesized caricature image revised by the image revision unit 15 or 55 is stored in the image storage unit 16 or 56, or displayed or printed by the output unit 17 or 57 in step S16.
The caricature image obtained through such a method can be immediately used for a particular purpose. Alternatively, the caricature image can be used as a draft when an artist produces a caricature, thereby increasing productivity during manual treatment.
A procedure of generating a caricature image using the sample image replacement method performed in the synthetic face image generating apparatuses 1 and 40 based on the shape information about a face image according to the present invention will be described with reference to FIG. 6.
FIG. 6 further includes step S35 and S36 in addition to steps shown in FIG. 5. Thus, descriptions of the other steps S30 through S34 and S37 through S39 will be omitted. In other words, the method of FIG. 6 is a method of replacing a part or the entire portion of a caricature image synthesized according to a method as shown in FIG. 5 with a sample image prepared in the sample image DB 333 or 633.
Sample images stored in the sample image DB 333 or 633 are formed based on the result of performing a statistic analysis on shape information about different faces. The sample image formation method can be divided into two cases: a case where transformation of a sample image is admitted, and a case where transformation of a sample image is not admitted.
According to the sample image formation method in the case where transformation of a sample image is admitted, a normalized sample image is formed, and the size and shape of the sample image
are transformed based on shape information S,„ extracted from an input face image in step S31. This method has advantages of substantially reflecting the shape of the input face image and requiring a small number of sample images but has disadvantages of distorting sample image by transformation and deteriorating entire picture quality.
When using the sample image formation method in the case where transformation of a sample image is not admitted, a part or the entire portion of a caricature image is replaced with only a previously formed sample image to synthesize a new caricature image. Accordingly, this method has an advantage of accomplishing high definition but is disadvantageous in that it is difficult to substantially reflect the shape of the input image and many sample images as possible as can be formed should be prepared in advance.
A method of measuring a difference D used in step S35 of FIG. 6 can be expressed by Equation (3).
D = W -∑(Ci, -Cn) + (l-W) -∑(Ctl -Cφ) ...(3)
;=1 (=1
Here, O≤ W≤ , Cs, (z-1 , ..., n) is shape information about an input image, CV(z'=1 , ..., n) is shape information about a sample image, Cu (z-1 , ..., n is the difference Tre/ between texture information about the input image and texture information about a reference image, and Cφ (z-1 , ..., ) is the difference Tref between texture information about the sample image and the texture information about the reference image. Instead of immediately applying shape information and texture information to Cs„ Crι, C.„ and Cq„ coefficients of eigenvectors obtained by linearly analyzing the shape information and the texture information, as shown in Equation (1), can be used. Here, each coefficient has a dimension of (mΛ).
The partial image replacing unit 323 or 623 measure the
difference D between the input image and each sample image in step S35 and replaces a part or the entire portion of a caricature image with a sample image having a minimum difference in step S36.
When a caricature image is generated according to the sample image replacement method, instead of the entire generated caricature image, the codes of a sample image with which a part or the entire portion of the caricature image will be replaced are compressed and transmitted in a low-speed communication environment, thereby markedly increasing a compression ratio. A procedure of designing a makeup performed in the synthetic face image generating apparatuses 1 and 40 based on the shape information about a face image according to the present invention will be described with reference to FIG. 7.
FIG. 7 is the same as FIG. 5, with the exception that FIG. 7 includes steps S42 through S45, instead of S22 through S24 of FIG. 5. Thus, descriptions of the other steps S40, S41 , and S46 through S48 will be omitted.
In step S42, the face image synthesizer 322 or 662 presents a variety of sample makeup images stored in the makeup image DB 334 or 634 to a user through the user interfacing device 10a, 10b or 50 so that the user can select a desired sample makeup image. Here, the sample makeup images have the same shape as that of a reference image.
In step S43, the face image synthesizer 322 or 622 transforms the selected sample makeup image or an image, which is generated by performing a weighted summation of the selected sample makeup image and an image reflecting texture information Tin extracted from an input face image in step S41 , using shape information Sin extracted from the input face image in step S41 , thereby synthesizing a face image in which the selected sample makeup image is applied to the user's face image. In step S44, the synthesized face image reflecting the sample
makeup image is transmitted to the user interfacing device 10a, 10b or 50 and displayed through the output unit 17 or 57, and the user command input unit 12 or 52 of the user interfacing device 10a, 10b or 50 receives a user control command instructing revision of the makeup of the displayed face image from the user. The face image synthesizer 322 or 622 revises the face image reflecting the sample makeup image according to the user control command, and the revised face image is transmitted to the user interfacing device 10a, 10b or 50 and displayed through the output unit 17 or 57. Thereafter, in step S45, the user command input unit 12 or 52 of the user interfacing device 10a, 10b or 50 receives from the user a user control command indicating whether the user is satisfied with the displayed face image. Here, when the user control command indicates "satisfied", the procedure proceeds to step S46 of adding an accessory. In contrast* when the user control command indicates "not satisfied", the procedure returns to step S42.
While this invention has been particularly shown and described with reference to preferred embodiments thereof, it will be understood by those skilled in the art that various changes in forms and details may be made therein without departing from the spirit and scope of the invention. The above embodiments have been used in a descriptive sense only and not for purpose of limitation. Therefore, the scope of the invention will be defined by the appended claims not by the above description, and it should be construed that modifications made within the scope of the invention are covered by the present invention.
Industrial Applicability
According to the present invention, first, shape information, which is represented by a deformation field with respect to a reference image, about an input face image is extracted from the input face image, and
images which have the same shape as that of the reference image but have different types of texture information and the extracted shape information about the input face image are used, thereby synthesizing natural new images reflecting the shape of the input face image and having high quality regardless of the state of the input face image. Accordingly, the present invention can be effectively utilized in a variety of fields such as generation of character images, virtual makeup design, making of montages for criminal search, animations and games.
Second, in case of generating caricature images, different caricatures including the characteristics of the shape of a user can be immediately generated, and a part or the entire portion of a generated caricature can be exaggerated or transformed. In addition, by using information about the shape of a user's face, a complex image revision process can be simplified and automatized, thereby improving productivity in generating characters.
Third, in case of makeup design, a user can easily create and check a makeup design on the user's face image and can easily revise the makeup design partially or entirely.
Fourth, the result of adding a variety of accessories to a new synthesized face image can be immediately checked. In addition, the present invention can be easily applied to a variety of applications such as realization of an avatar in the virtual reality based on shape information, restoration of a three-dimensional face image, and video chatting which need a face image.