WO2007118919A1 - Method for generating synthetic-animation images - Google Patents

Method for generating synthetic-animation images Download PDF

Info

Publication number
WO2007118919A1
WO2007118919A1 PCT/ES2007/000235 ES2007000235W WO2007118919A1 WO 2007118919 A1 WO2007118919 A1 WO 2007118919A1 ES 2007000235 W ES2007000235 W ES 2007000235W WO 2007118919 A1 WO2007118919 A1 WO 2007118919A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
model
images
dimensional
image
Prior art date
Application number
PCT/ES2007/000235
Other languages
Spanish (es)
French (fr)
Inventor
Juan Coll Soler
Ãlvaro UÑA RESA
Original Assignee
Emotique, S.L.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Emotique, S.L. filed Critical Emotique, S.L.
Publication of WO2007118919A1 publication Critical patent/WO2007118919A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • G06T13/203D [Three Dimensional] animation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects

Definitions

  • the present invention relates to a method for generating synthetic animation * images for the leisure sector in general, such as the film industry, videogames and promotional advertising among others, in which the personalization of an image or animation is of interest. kinematics.
  • the procedure for the generation of synthetic animation images of this invention presents some technical peculiarities that allow the realization of audiovisual means of synthetic image that allow one or more users to personalize them to be able to configure one or more characters of the story or event described in the audiovisual medium with the face and / or body of said users.
  • the procedure seeks to obtain a method or system that allows a personalization of, for example, a film, with the participation of the user, whose image and similarity is easily incorporated into one of the characters.
  • the procedure also allows the creation of videogames and interactive software in which the user can also configure one or more characters to his image and similarity, allowing the hardware of the machine that performs the image synthesis a virtual representation of the user in the graphic environment of said game.
  • the procedure has several phases that include:
  • This procedure is integrated so that the user can configure the final result very easily and easily.
  • a number of important steps have been provided in the procedure that allow considerable possibilities of interacting and modifying the work parameters, depending on the object to be obtained by executing the procedure.
  • the acquisition of at least one image of the user is carried out by means of a video camera, photography or the like, such as a "webcam” or a mobile camera.
  • a video camera photography or the like, such as a "webcam” or a mobile camera.
  • These images comprise the face of the user seen preferably from the front, this image being configured as a two-dimensional digital dot map.
  • the process of recognizing the user's facial and / or body features can be carried out by means of suitable software that executes an artificial intelligence algorithm applied to the artificial vision, already existing in the market, from said image in static format if the quality It is enough for the correct contrast.
  • an artificial intelligence algorithm applied to the artificial vision, already existing in the market, from said image in static format if the quality It is enough for the correct contrast.
  • the camera is connected to the computer that performs said recognition in real time, it allows the recognition to be performed automatically in real time. In the cases mentioned above, this recognition is executed automatically without the user having to handle any part of the algorithm.
  • the objective of this algorithm is to automate the process of distinguishing between the various objects that they may appear in the image, discarding those that do not work and focusing on the only object of interest, which is the user's face.
  • the facial features of this face are identified, such as, for example, the position and size of the eyes, mouth, nose, and others.
  • a contour pattern of the face is used, which comprises, for example, the eyebrows, the eyes, the nose, the lips, the jaw and the cheekbones, among others.
  • This pattern adapts to the image automatically by the software, obtaining a personalized shift of key coordinates with respect to a neutral pattern. This displacement corresponds to the physiognomic characteristics of the user.
  • it is possible to alter a model, for example of the head to adapt it to the physiognomy of the user and thus be able to superimpose the image as texture with complete accuracy. It is possible that the model may have a beard or mustache if they have been detected in the recognition of the image.
  • the modification can be made in real time.
  • the generated custom model can be a single portion of the character, such as the head, so that the generation of the complete three-dimensional model includes the use of portions of pre-established models, In addition to the custom models generated.
  • An example is the coupling of a custom head model of the user to an already designed comic body. This hybrid model is fully manageable in the synthesis of modeling images.
  • This synthesis or "rendering" of the customized models, whether complete or hybrid, is performed in a virtual scenario or background, together with the pre-established models that are necessary to obtain the sequence of images that gives rise to the animation.
  • the synthesization can be configured passively, that is, the production software generates a movie or sequence according to the script previously established for the animation and can be turned into a physical medium for later viewing or sent to a display device, such as a television or analog screen, as a cinematic medium, obtaining one or several movie sequences with at least one custom character.
  • This synthesis of images can occur in real time, so that the user is allowed to intervene in the scene and manage his character, a method especially suitable for video games.
  • the user can have adequate control means for the interaction and alteration in real time of the characteristics of the customized model.
  • These interactions may correspond to gestures and actions that represent facial emotions, such as laughing, crying or showing anger or body movements, such as speaking, opening and closing the eyes or body movements.
  • the image dump of the successive frames obtained in the synthesis of images can be sent to various devices.
  • digital files can be generated by way of cinematic sequences to be dumped into physical media (DVD or others) or sent to multimedia devices such as mobile phones.
  • the animation produced can be viewed in real time, for example on a television screen, cinematographic or broadcast on the Internet, interconnected computer networks or broadcasting. Description of the figures.
  • FIG. 1 shows a process flow scheme
  • Figure 2 shows a diagram of the phases of capture and recognition of facial features from the image obtained.
  • Figure 3 shows a diagram of the phases of generation of the corresponding mathematical three-dimensional model with the recognized features of the user.
  • Figure 4 shows a scheme of the synthesis of a scene in which a character has been introduced with the custom model added and a second character consisting of a preset model, with user control means for interactive management. representative model of his character.
  • the procedure comprises the following sequence of phases:
  • This software module comprises an artificial intelligence module applied to the artificial vision, which makes a recognition (2) of the facial features from a neutral pattern (21) that adapts to a custom pattern (14) as they are the characteristics of the captured image (12) of the user.
  • Another software module evaluates the custom pattern (14) obtained and generates the three-dimensional mesh model (11) corresponding to the user's features, from the deformation of a pre-established neutral three-dimensional mesh model (31). In this generation, the surface textures obtained from the image (12) or acquired images are superimposed on said model (11). External modifiers (32) are also included, such as the design of glasses, for example, or physiognomic modifications, such as the modification of pointed ears among others.
  • the custom head model (11) is coupled to a portion of the preset body model (42), constituting a mixed three-dimensional model (6) and introduced into a virtual scenario (43) together with other preset models (41 ).
  • control means (7) such as game controls, for the indirect handling of points and curves of the custom model (6, 11 ) for the representation of facial gestures and emotions.
  • One dump (5) is produced per output device of the consecutive frames generated by a synthesizing hardware for display on television screen. It is also possible to proceed to the recording of said frames on physical media, such as DVD or other.

Abstract

Method that comprises the phases of: acquisition (1) of at least one frontal image (12) of the user - user's face and/or full body - preferably by means of a video or stills camera (13); recognition (2) of the user's facial and/or body features on the basis of the acquired images (12); generation (3) of a three-dimensional grid model (11) corresponding to the user's features, comprising the deformation of a pre-established neutral three-dimensional grid model (31), according to the features recognised and the superimposition of the images (12) in accordance with the modified structure; introduction of the model (11) into a virtual scenario (43), with other, personalized or pre-established models (41); synthesis (4) of modelling images, with or without interaction of the personalized models (11), pre-established models (41) and scenario (43) for display thereof; dumping (5) of the successive photograms obtained in the image synthesis (4) by an output device.

Description

DESCRIPCIÓN DESCRIPTION
PROCEDIMIENTO PARA LA GENERACIÓN DE IMÁGENES DE ANIMACIÓNPROCEDURE FOR THE GENERATION OF ANIMATION IMAGES
SINTÉTICA.SYNTHETIC
Objeto de Ia invenciónObject of the invention
La presente invención se refiere a un procedimiento para Ia generación de imágenes de animación* sintética para el sector del ocio en general, tal como Ia industria cinematográfica, los videojuegos y publicidad promocional entre otros, en el que interese Ia personalización de una imagen o animación cinemática.The present invention relates to a method for generating synthetic animation * images for the leisure sector in general, such as the film industry, videogames and promotional advertising among others, in which the personalization of an image or animation is of interest. kinematics.
Antecedentes de Ia invención.Background of the invention.
En Ia actualidad el uso de ordenadores para Ia generación de imágenes de animación sintética se ha extendido notablemente en el sector del ocio.At present, the use of computers for the generation of synthetic animation images has extended notably in the leisure sector.
Tanto en Ia cinematografía como en el mundo de los videojuegos ya es común Ia generación de personajes y escenarios visuales a partir de complejos modelos matemáticos y representaciones virtuales.Both in the cinematography and in the world of videogames, the generation of characters and visual scenarios from complex mathematical models and virtual representations is already common.
En el caso de Ia cinematografía se han realizado sustituciones de personajes y escenarios que se han generado por ordenador mediante herramientas de software que recrean espacios y objetos tridimensionales, los cuales mediante el montaje adecuado se pueden superponer con las imágenes previamente grabadas de los actores convencionales. Un ejemplo de ellos son las películas "La guerra de las galaxias" de Lucasfilm o "El señor de los anillos" de Wingnut Films. La calidad de imagen obtenida puede ser de tal consideración que incluso no es posible diferenciar el personaje generado de una persona normal si no es por sus características fantásticas.In the case of cinematography, substitutions of characters and scenarios that have been generated by computer have been made using software tools that recreate three-dimensional spaces and objects, which by means of proper assembly can be superimposed with the previously recorded images of conventional actors. An example of them are the films "Star Wars" by Lucasfilm or "The Lord of the Rings" by Wingnut Films. The image quality obtained can be of such consideration that it is not even possible to differentiate the generated character from a normal person if it is not for their fantastic characteristics.
Estas técnicas permiten que sobre una imagen cualquiera se realice con una cierta facilidad Ia superposición de imágenes de personas reales o personajes generados en imágenes ya grabadas. Ia depuración en el uso de esta técnica ha llevado a Ia creación de películas realizadas totalmente en imagen sintética, por ejemplo "Shreck" de Dreamworks Pictures, en Ia que los escenario y los personajes son modelos tridimensionales preestablecidos, quedando como única participación para el ser humano el doblaje de las voces. El principal inconveniente es que estas técnicas requieren un personal de trabajo especializado muy entrenado y gran cantidad de recursos informáticos. Otro inconveniente es que resulta una experiencia pasiva en Ia que el usuario o espectador no puede intervenir.These techniques allow the overlapping of images of real people or characters generated in already recorded images on a single image. The purification in the use of this technique has led to the creation of films made entirely in synthetic image, for example "Shreck" of Dreamworks Pictures, in which the stage and the characters are pre-established three-dimensional models, leaving the voice dubbing as the only participation for the human being. The main drawback is that these techniques require highly trained specialized work personnel and a large number of computer resources. Another drawback is that it is a passive experience in which the user or spectator cannot intervene.
Estas técnicas se han tratado de utilizar en publicidad por Internet, permitiendo insertar una fotografía de un usuario, debidamente recortada, en una secuencia ya pregrabada, de forma que proporcione Ia sensación de que el usuario participa en dicha secuencia, consiguiendo así un importante impacto promocional y atractivo del posible consumidor. Estas técnicas están limitadas al hecho de que son imágenes bidimensionales, tanto Ia fotografía del usuario como Ia secuencia pregrabada, con Io que Ia sensación de "parche" es visible de forma notoria, sin poder girar Ia cara personalizada y realizar pequeñas modificaciones de Ia articulación de Ia boca mediante sistemas de deformación de Ia imagen (conocidos como "morphing") o sustitución por capas (sistema "Flash" de Macromedia), con resultados un tanto bruscos.These techniques have been tried to use in Internet advertising, allowing to insert a photograph of a user, duly cropped, in a pre-recorded sequence, so as to provide the feeling that the user participates in said sequence, thus achieving an important promotional impact and attractiveness of the potential consumer. These techniques are limited to the fact that they are two-dimensional images, both the photograph of the user and the pre-recorded sequence, so that the feeling of "patch" is visible in a visible way, without being able to rotate the personalized face and make small modifications of the joint of the mouth by means of deformation systems of the image (known as "morphing") or substitution by layers (Macromedia's "Flash" system), with somewhat abrupt results.
A su vez, en el sector de los videojuegos se han extendido muchos títulos basados en entornos tridimensionales, que posteriormente son sintetizados o "renderizados" por el hardware de Ia consola de videojuegos para su visualización en pantallas de televisión en tiempo real. Estos juegos comprenden en Ia estructura de su programación unos personajes u objetos previamente realizados en modelos preestablecidos y unos escenarios también modelados tridimensionalmente de forma preestablecida para su interacción. El jugador mediante los mandos de control adecuados puede manipular uno o varios personajes protagonistas y navega en los escenarios generados según sea el guión previsto del juego, con las correspondientes libertades que se Ie hayan asignado. La sintetización de imágenes desde los modelos 3D disponibles no alcanzan Ia calidad de imagen cinematográfica, pero se Ie acercan Io suficiente como para que en una pantalla de televisión resulten sustancialmente creíbles y convincentes. Sin embargo las posibilidades de personalización del personaje por parte del usuario pasan por Ia asignación de un nombre y Ia configuración según unas opciones ya preestablecidas, con Io que el personaje protagonista del juego no resulta más que una combinación de dichas opciones, sin posibilidad de realizar una personalización auténtica.In turn, in the videogames sector, many titles based on three-dimensional environments have been extended, which are subsequently synthesized or "rendered" by the hardware of the video game console for display on real-time television screens. These games include in the structure of their programming characters or objects previously made in pre-established models and scenarios also modeled three-dimensionally in a pre-established way for their interaction. The player through the appropriate control commands can manipulate one or several main characters and navigate the scenarios generated according to the intended script of the game, with the corresponding freedoms that have been assigned. The synthesization of images from the available 3D models does not reach the cinematographic image quality, but they are close enough that they are substantially credible and convincing on a television screen. However, the possibilities of personalization of the character by the user go through the assignment of a name and the configuration according to pre-established options, so that the main character of the game is not more than a combination of these options, without the possibility of performing a authentic customization.
Dentro de Ia industria de Ia infografía se han tratado de realizar varios intentos de realizar modelos de cabezas y bustos tridimensionales que pudieran representar las facciones de un usuario. Su uso está más orientado a los denominados locutores virtuales. Un ejemplo es el software "3DMeNow!" de Ia empresa BioVirtual, Inc. que a partir de dos fotografías, una de frente y otra de perfil de Ia cabeza del usuario permite Ia reconstrucción de un modelo virtual de su cabeza con una semejanza considerable. En este software el usuario debe introducir ambas imágenes anteriormente citadas y utilizando unas guías de puntos determinar que zonas de las imágenes corresponden con puntos críticos, tal como son las cejas, los ojos o Ia boca. Este proceso requiere una habilidad de manejo notoria y una cierta dedicación por parte del usuario, para obtener un resultado relativamente bueno.Within the infographic industry, several attempts have been made to make models of three-dimensional heads and busts that could represent the factions of a user. Its use is more oriented to the so-called virtual speakers. An example is the software "3DMeNow!" of the company BioVirtual, Inc. that from two photographs, one in front and the other in the profile of the user's head allows the reconstruction of a virtual model of his head with a considerable similarity. In this software, the user must enter both images mentioned above and using point guides determine which areas of the images correspond to critical points, such as eyebrows, eyes or mouth. This process requires a remarkable handling ability and a certain dedication on the part of the user, to obtain a relatively good result.
También existen otros programas que permiten Ia superposición de una imagen recortada de Ia cara directamente sobre un modelo tridimensional de una cabeza neutra. Este procedimiento, aunque sencillo, tampoco produce efectos de calidad suficiente, ya que Ia morfología de Ia cabeza está preestablecida y Ia disparidad de Ia deformación de Ia imagen bidimensional sobre Ia morfología original se observa notoriamente, mediante una deformación que puede hacer irreconocible al usuario en el modelo visualizado, un ejemplo de programa que realiza esta función es "Poser" de Curious Labs, Inc.There are also other programs that allow the superposition of a cropped image of the face directly on a three-dimensional model of a neutral head. This procedure, although simple, also does not produce effects of sufficient quality, since the morphology of the head is preset and the disparity of the deformation of the two-dimensional image on the original morphology is notoriously observed, by means of a deformation that can make the user unrecognizable in the model displayed, an example of a program that performs this function is "Poser" by Curious Labs, Inc.
Descripción de Ia invenciónDescription of the invention
El procedimiento para Ia generación de imágenes de animación sintética de esta invención presenta unas particularidades técnicas que permiten Ia realización de medios audiovisuales de imagen sintética que permiten a uno o varios usuarios Ia personalización de los mismos al poder configurar uno o más personajes de Ia historia o suceso descrito en el medio audiovisual con Ia cara y/o cuerpo de dichos usuarios.The procedure for the generation of synthetic animation images of this invention presents some technical peculiarities that allow the realization of audiovisual means of synthetic image that allow one or more users to personalize them to be able to configure one or more characters of the story or event described in the audiovisual medium with the face and / or body of said users.
En efecto, el procedimiento busca obtener un método o sistema que permita una personalización de, por ejemplo una película, con Ia participación del usuario, cuya imagen y semejanza se incorpora en uno de los personaje fácilmente. El - A -In effect, the procedure seeks to obtain a method or system that allows a personalization of, for example, a film, with the participation of the user, whose image and similarity is easily incorporated into one of the characters. He - TO -
procedimiento también permite Ia creación de videojuegos y software interactivo en el que el usuario puede configurar igualmente uno o más personajes a su imagen y semejanza, permitiendo el hardware de Ia máquina que realiza Ia síntesis de imagen una representación virtual del usuario en el entorno gráfico de dicho juego.The procedure also allows the creation of videogames and interactive software in which the user can also configure one or more characters to his image and similarity, allowing the hardware of the machine that performs the image synthesis a virtual representation of the user in the graphic environment of said game.
El procedimiento presenta varias fases que comprenden:The procedure has several phases that include:
- Adquisición de al menos una imagen del usuario.- Acquisition of at least one user image.
- Reconocimiento de los rasgos faciales y/o corporales del usuario - Generación de un modelo tridimensional matemático personalizado.- Recognition of the user's facial and / or body features - Generation of a personalized mathematical three-dimensional model.
- Introducción del modelo personalizado en un escenario virtual.- Introduction of the customized model in a virtual scenario.
- Sintetización de las imágenes de modelado, volcado por dispositivo de salida.- Synthesis of modeling images, dump by output device.
Este procedimiento está integrado de forma que el usuario pueda configurar el resultado final de forma muy fácil y sencilla. Sin embargo, se han previsto en el procedimiento un número de pasos importantes que permiten unas posibilidades de interactuar y modificar los parámetros de trabajo considerables, dependiendo el objeto que se desea obtener mediante Ia ejecución del procedimiento.This procedure is integrated so that the user can configure the final result very easily and easily. However, a number of important steps have been provided in the procedure that allow considerable possibilities of interacting and modifying the work parameters, depending on the object to be obtained by executing the procedure.
La adquisición de al menos una imagen del usuario se realiza mediante una cámara de vídeo, fotografía o análoga, tal como una "webcam" o Ia cámara de un móvil. Estas imágenes comprenden el rostro del usuario visto de frente preferentemente, estando esta imagen configurada como un mapa de puntos digital bidimensional.The acquisition of at least one image of the user is carried out by means of a video camera, photography or the like, such as a "webcam" or a mobile camera. These images comprise the face of the user seen preferably from the front, this image being configured as a two-dimensional digital dot map.
El proceso de reconocimiento de los rasgos faciales y/o corporales del usuario se puede realizar mediante software adecuado que ejecuta un algoritmo de inteligencia artificial aplicado a Ia visión artificial, ya existente en el mercado, a partir de dicha imagen en formato estático si Ia calidad es suficiente para el correcto contraste. Aunque si Ia cámara esta conectada con el ordenador que realiza dicho reconocimiento en tiempo real, permite que el reconocimiento se realice de forma automática en tiempo real. En los casos anteriormente citados este reconocimiento se ejecuta de forma automática sin que el usuario deba manejar parte alguna del algoritmo. El objetivo de dicho algoritmo es automatizar el proceso de distinguir entre los diversos objetos que puedan aparecer en Ia imagen, descartando los que no sirven y centrándose en el único objeto que interesa, que es Ia cara del usuario. Una vez identificada Ia región en Ia imagen que ocupa Ia cara, se procede a identificar los rasgos faciales de esta, como pueden ser, por ejemplo Ia posición y tamaño de los ojos, boca, nariz, y otros. Durante dicho reconocimiento se utiliza un patrón de contorno de Ia cara que comprende, por ejemplo, las cejas, los ojos, Ia nariz, los labios, mandíbula y pómulos entre otros. Este patrón se acomoda a Ia imagen de forma automática por el software obteniéndose un desplazamiento personalizado de coordenadas clave respecto a un patrón neutro. Este desplazamiento se corresponde con las características fisonómicas del usuario. De esta forma con el patrón modificado por el reconocimiento y Ia imagen es posible alterar un modelo, por ejemplo de Ia cabeza, para adaptarlo a Ia fisonomía del usuario y así poder superponer Ia imagen como textura con una completa exactitud. Es posible que el modelo pueda presentar barba o bigote si se han detectado en el reconocimiento de Ia imagen.The process of recognizing the user's facial and / or body features can be carried out by means of suitable software that executes an artificial intelligence algorithm applied to the artificial vision, already existing in the market, from said image in static format if the quality It is enough for the correct contrast. Although if the camera is connected to the computer that performs said recognition in real time, it allows the recognition to be performed automatically in real time. In the cases mentioned above, this recognition is executed automatically without the user having to handle any part of the algorithm. The objective of this algorithm is to automate the process of distinguishing between the various objects that they may appear in the image, discarding those that do not work and focusing on the only object of interest, which is the user's face. Once the region in the image that occupies the face is identified, the facial features of this face are identified, such as, for example, the position and size of the eyes, mouth, nose, and others. During said recognition, a contour pattern of the face is used, which comprises, for example, the eyebrows, the eyes, the nose, the lips, the jaw and the cheekbones, among others. This pattern adapts to the image automatically by the software, obtaining a personalized shift of key coordinates with respect to a neutral pattern. This displacement corresponds to the physiognomic characteristics of the user. In this way, with the pattern modified by the recognition and the image, it is possible to alter a model, for example of the head, to adapt it to the physiognomy of the user and thus be able to superimpose the image as texture with complete accuracy. It is possible that the model may have a beard or mustache if they have been detected in the recognition of the image.
Con los datos obtenidos a través del reconocimiento de visión artificial, se procede a deformar una malla tridimensional de un modelo neutro de cabeza, de modo que esta malla acabe ajustándose a las características fisonómicas del usuario. Posteriormente se aplica Ia imagen del usuario encima de este modelo personalizado de malla, que quedará perfectamente acomodada, reflejando con exactitud Ia apariencia del usuario.With the data obtained through the recognition of artificial vision, a three-dimensional mesh of a neutral head model is deformed, so that this mesh ends up adjusting to the physiognomic characteristics of the user. Subsequently, the image of the user is applied on top of this custom mesh model, which will be perfectly accommodated, accurately reflecting the user's appearance.
En este modelo es posible realizar modificaciones fisonómicas, mediante modificadores externos y readaptar según sean los deseos del usuario o los requerimientos del entorno o guión en el que desarrolla Ia acción cinemática, alterando su representación. Estas modificaciones pueden ser variadas, tal como añadir unas gafas, accesorios o cambiar el tipo de peinado, entre otros.In this model it is possible to make physiognomic modifications, by means of external modifiers and to adapt according to the wishes of the user or the requirements of the environment or script in which he develops the kinematic action, altering its representation. These modifications can be varied, such as adding some glasses, accessories or changing the type of hairstyle, among others.
Igualmente que el reconocimiento en el caso de que Ia cámara esté conectada directamente con el ordenador o hardware que ejecuta los algoritmos, Ia modificación se puede realizar en tiempo real.Like the recognition in the case that the camera is directly connected to the computer or hardware that executes the algorithms, the modification can be made in real time.
Tal como se ha comentado, el modelo personalizado generado puede ser una sola porción del personaje, tal como Ia cabeza, por Io que Ia generación del modelo tridimensional completo comprende el uso de porciones de modelos preestablecidas, además de los modelos personalizados generados. Un ejemplo es el acoplamiento de un modelo de cabeza personalizada del usuario a un cuerpo cómico ya diseñado. Este modelo híbrido es totalmente manejable en Ia sintetización de las imágenes de modelado.As mentioned, the generated custom model can be a single portion of the character, such as the head, so that the generation of the complete three-dimensional model includes the use of portions of pre-established models, In addition to the custom models generated. An example is the coupling of a custom head model of the user to an already designed comic body. This hybrid model is fully manageable in the synthesis of modeling images.
Esta sintetización o "renderizado" de los modelos personalizados, ya sean completos o híbridos se realiza en un escenario virtual o fondo, junto con los modelos preestablecidos que se crean necesarios para obtener Ia secuencia de imágenes que da lugar a Ia animación. La sintetización puede configurarse de forma pasiva, es decir, el software de producción genera una película o secuencia según el guión previamente establecido para Ia animación y se puede volcar en un medio físico para su visionado posterior o se envía a un dispositivo de visualización, tal como una pantalla de televisión o análoga, a modo de medio cinematográfico, obteniéndose una o varias secuencias de película con al menos un personaje personalizado.This synthesis or "rendering" of the customized models, whether complete or hybrid, is performed in a virtual scenario or background, together with the pre-established models that are necessary to obtain the sequence of images that gives rise to the animation. The synthesization can be configured passively, that is, the production software generates a movie or sequence according to the script previously established for the animation and can be turned into a physical medium for later viewing or sent to a display device, such as a television or analog screen, as a cinematic medium, obtaining one or several movie sequences with at least one custom character.
Esta sintetización de imágenes puede producirse en tiempo real, con Io que se permite que el usuario pueda intervenir en Ia escena y maneje su personaje, método especialmente adecuado para los videojuegos. Así el usuario puede disponer de unos medios de control adecuados para Ia interactuación y alteración en tiempo real de las características del modelo personalizado. Estas interactuación puede corresponder a gestos y acciones que representan emociones faciales, tales como reír, llorar o mostrar enfado o movimientos corporales, tales como hablar, abrir y cerrar los ojos o movimientos del cuerpo.This synthesis of images can occur in real time, so that the user is allowed to intervene in the scene and manage his character, a method especially suitable for video games. Thus, the user can have adequate control means for the interaction and alteration in real time of the characteristics of the customized model. These interactions may correspond to gestures and actions that represent facial emotions, such as laughing, crying or showing anger or body movements, such as speaking, opening and closing the eyes or body movements.
Tal como se ha mencionado el volcado de la imagen de los sucesivos fotogramas obtenidos en Ia sintetización de imágenes se pueden enviar a diversos dispositivos. Por ejemplo se pueden generar ficheros digitales a modo de secuencias cinematográficas para su volcado en soporte físico (DVD u otros) o envío a dispositivos multimedia tal como los teléfonos móviles. En otro sentido Ia animación producida se puede visionar en tiempo real, por ejemplo en una pantalla de televisión, cinematográfica o difundida por Internet, redes de ordenadores interconectados o radiodifusión. Descripción de las figuras.As mentioned, the image dump of the successive frames obtained in the synthesis of images can be sent to various devices. For example, digital files can be generated by way of cinematic sequences to be dumped into physical media (DVD or others) or sent to multimedia devices such as mobile phones. In another sense, the animation produced can be viewed in real time, for example on a television screen, cinematographic or broadcast on the Internet, interconnected computer networks or broadcasting. Description of the figures.
Para complementar Ia descripción que se está realizando y con objeto de facilitar Ia comprensión de las características de Ia invención, se acompaña a Ia presente memoria descriptiva un juego de dibujos en los que, con carácter ilustrativo y no limitativo, se ha representado Io siguiente:To complement the description that is being made and in order to facilitate the understanding of the characteristics of the invention, a set of drawings is attached to the present specification in which, for illustrative and non-limiting purposes, the following has been represented:
- La figura 1 muestra un esquema de flujo del proceso.- Figure 1 shows a process flow scheme.
La figura 2 muestra un esquema de las fases de captura y reconocimiento de los rasgos faciales a partir de Ia imagen obtenida.Figure 2 shows a diagram of the phases of capture and recognition of facial features from the image obtained.
La figura 3 muestra un esquema de las fases de generación del modelo tridimensional matemático correspondiente con los rasgos reconocidos del usuario.Figure 3 shows a diagram of the phases of generation of the corresponding mathematical three-dimensional model with the recognized features of the user.
La figura 4 muestra un esquema de Ia sintetización de una escena en Ia que se ha introducido un personaje con el modelo personalizado añadido y un segundo personaje constituido por un modelo preestablecido, contando con unos medios de control por parte del usuario para el manejo interactivo el modelo representante de su personaje.Figure 4 shows a scheme of the synthesis of a scene in which a character has been introduced with the custom model added and a second character consisting of a preset model, with user control means for interactive management. representative model of his character.
Realización preferente de Ia invenciónPreferred embodiment of the invention
Como se puede observar en las figuras referenciadas para Ia generación de una secuencia animada en Ia que participa un modelo tridimensional (6) de un personaje que comprende un modelo tridimensional de cabeza (11) modelada a partir de Ia imagen (12) del usuario y otro personaje conformado a partir de un modelo tridimensional (41) preestablecido para su interacción según un guión de resolución variable, el procedimiento comprende Ia siguiente secuencia de fases:As can be seen in the referenced figures for the generation of an animated sequence in which a three-dimensional model (6) of a character comprising a three-dimensional head model (11) modeled from the image (12) of the user and Another character formed from a three-dimensional model (41) preset for interaction according to a variable resolution script, the procedure comprises the following sequence of phases:
Adquisición (1) mediante una cámara de video (13) de una imagen frontalAcquisition (1) by a video camera (13) of a front image
(12) del rostro del usuario, estando esta cámara de video (13) conectada con un ordenador en el que se ejecuta un programa o módulo de software de captura de dicha imagen. Este módulo de software comprende un módulo de inteligencia artificial aplicada a Ia visión artificial, que efectúa un reconocimiento (2) de los rasgos faciales a partir de un patrón neutro (21) que adapta a un patrón personalizado (14) según sean las características de Ia imagen capturada (12) del usuario.(12) of the user's face, this video camera (13) being connected to a computer on which a program or software module for capturing said image is executed. This software module comprises an artificial intelligence module applied to the artificial vision, which makes a recognition (2) of the facial features from a neutral pattern (21) that adapts to a custom pattern (14) as they are the characteristics of the captured image (12) of the user.
Otro módulo de software evalúa el patrón personalizado (14) obtenido y genera el modelo de malla tridimensional (11) correspondiente con los rasgos del usuario, a partir de Ia deformación de un modelo de malla tridimensional neutro (31) preestablecido. En esta generación se superponen sobre dicho modelo (11) las texturas de superficie obtenidas a partir de Ia imagen (12) o imágenes adquiridas. También se incluyen unos modificadores externos (32), tales como el diseño de gafas, por ejemplo, o modificaciones fisonómicas, tales como Ia modificación de orejas puntiagudas entre otros. En una siguiente fase el modelo personalizado (11) de cabeza es acoplado a una porción de modelo preestablecido (42) de cuerpo, constituyendo un modelo tridimensional mixto (6) e introducido en un escenario virtual (43) junto con otros modelos preestablecido (41).Another software module evaluates the custom pattern (14) obtained and generates the three-dimensional mesh model (11) corresponding to the user's features, from the deformation of a pre-established neutral three-dimensional mesh model (31). In this generation, the surface textures obtained from the image (12) or acquired images are superimposed on said model (11). External modifiers (32) are also included, such as the design of glasses, for example, or physiognomic modifications, such as the modification of pointed ears among others. In a next phase the custom head model (11) is coupled to a portion of the preset body model (42), constituting a mixed three-dimensional model (6) and introduced into a virtual scenario (43) together with other preset models (41 ).
Se procede a Ia sintetización (4) de imágenes de modelado con Ia interactuación del usuario a través de unos medios de control (7), tal como unos mandos de juego, para el manejo indirecto de puntos y curvas del modelo personalizado (6, 11) para Ia representación de gestos faciales y emociones.We proceed to the synthesis (4) of modeling images with the user's interaction through control means (7), such as game controls, for the indirect handling of points and curves of the custom model (6, 11 ) for the representation of facial gestures and emotions.
Se produce un volcado (5) por dispositivo de salida de los consecutivos fotogramas generados por un hardware de sintetización para su visualización en pantalla de televisión. También puede procederse a Ia grabación de dichos fotogramas en soporte físico, tal como DVD u otro.One dump (5) is produced per output device of the consecutive frames generated by a synthesizing hardware for display on television screen. It is also possible to proceed to the recording of said frames on physical media, such as DVD or other.
Una vez descrita suficientemente Ia naturaleza de Ia invención, así como un ejemplo de realización preferente, se hace constar a los efectos oportunos que los materiales, forma, tamaño y disposición de los elementos descritos podrán ser modificados, siempre y cuando ello no suponga una alteración de las características esenciales de Ia invención que se reivindican a continuación. Once the nature of the invention has been sufficiently described, as well as an example of a preferred embodiment, it is stated for the appropriate purposes that the materials, shape, size and arrangement of the described elements may be modified, provided that this does not imply an alteration of the essential characteristics of the invention that are claimed below.

Claims

REIVINDICACIONES
1.- Procedimiento para Ia generación de imágenes de animación sintética en1.- Procedure for the generation of synthetic animation images in
Ia que se usan modelos y escenarios preestablecidos caracterizado porque comprende las fases de: - adquisición (1) de al menos una imagen frontal (12) del usuario, de su rostro y/o del cuerpo entero, reconocimiento (2) de los rasgos faciales y/o corporales del usuario a partir de Ia imagen o imágenes adquiridas (12), generación (3) de un modelo de malla tridimensional (11) correspondiente con los rasgos del usuario, comprendiendo Ia deformación de un modelo de malla tridimensional neutro preestablecido (31) según los rasgos reconocidos y Ia superposición sobre dicho modelo de malla tridimensional (11) de las texturas obtenidas de las imágenes adquiridas (12), - introducción del modelo tridimensional (11) en un escenario virtualIa which preset models and scenarios are used characterized in that it comprises the phases of: - acquisition (1) of at least one frontal image (12) of the user, of his face and / or of the entire body, recognition (2) of the facial features and / or body of the user from the image or images acquired (12), generation (3) of a three-dimensional mesh model (11) corresponding to the user's features, comprising the deformation of a pre-established neutral three-dimensional mesh model ( 31) according to the recognized features and the superposition on said three-dimensional mesh model (11) of the textures obtained from the acquired images (12), - introduction of the three-dimensional model (11) in a virtual scenario
(43) con otros modelos personalizados o preestablecidos (41), sintetización (4) de imágenes de modelado con o sin interactuación de los modelos tridimensionales personalizados (11), modelos preestablecidos (41) y escenario (43) para su visualización. - volcado (5) por dispositivo de salida de los sucesivos fotogramas obtenidos en Ia sintetización (4) de imágenes.(43) with other custom or preset models (41), synthesization (4) of modeling images with or without interaction of the custom three-dimensional models (11), preset models (41) and stage (43) for viewing. - Dump (5) by output device of the successive frames obtained in the synthesis (4) of images.
2.- Procedimiento, según Ia reivindicación 1 , caracterizado porque Ia adquisición (1) de Ia imagen del usuario (12) es mediante una cámara de video (13), fotografía o análoga que permite Ia digitalización en mapa de puntos bidimensional.2. Method according to claim 1, characterized in that the acquisition (1) of the image of the user (12) is by means of a video camera (13), photograph or analogue that allows the two-dimensional point map scanning.
3.- Procedimiento, según Ia reivindicación 1 , caracterizado porque Ia adquisición (1) y reconocimiento (2) de los rasgos faciales es mediante medios de un algoritmo de inteligencia artificial aplicado a Ia visión artificial.3. Procedure according to claim 1, characterized in that the acquisition (1) and recognition (2) of facial features is by means of an artificial intelligence algorithm applied to artificial vision.
4.- Procedimiento, según Ia reivindicación 1 , caracterizado porque en Ia generación (3) del modelo tridimensional se introducen modificaciones de dicho modelo (11) mediante modificadores (32) para alterar su representación.4. Method according to claim 1, characterized in that in the generation (3) of the three-dimensional model modifications of said model (11) are introduced by modifiers (32) to alter its representation.
5.- Procedimiento, según Ia reivindicación 1 , caracterizado porque la generación (3) del modelo tridimensional (6) comprende el uso de porciones de modelo preestablecidas (42) y porciones de modelos personalizados generados (11) para Ia formación de un modelo utilizable en el escenario virtual (43).5. Procedure according to claim 1, characterized in that the generation (3) of the three-dimensional model (6) comprises the use of pre-established model portions (42) and portions of custom generated models (11) for the formation of a usable model in the virtual scenario (43).
6.- Procedimiento, según Ia reivindicación 1 , caracterizado porque Ia sintetización (4) de imágenes de modelado es en tiempo real.6. Method according to claim 1, characterized in that the synthesis (4) of modeling images is in real time.
7.- Procedimiento, según cualquiera de las reivindicaciones 1 y 6, caracterizado porque comprende Ia interactuación del usuario mediante medios de control (7) adecuados para Ia alteración en tiempo real de las características del modelo personalizado (6, 11) correspondientes a gestos y acciones que representan emociones faciales o movimientos corporales. 7. Method according to any of claims 1 and 6, characterized in that it comprises user interaction by means of control (7) suitable for the real-time alteration of the characteristics of the custom model (6, 11) corresponding to gestures and actions that represent facial emotions or body movements.
PCT/ES2007/000235 2006-04-19 2007-04-19 Method for generating synthetic-animation images WO2007118919A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
ESP200600993 2006-04-19
ES200600993A ES2284391B1 (en) 2006-04-19 2006-04-19 PROCEDURE FOR THE GENERATION OF SYNTHETIC ANIMATION IMAGES.

Publications (1)

Publication Number Publication Date
WO2007118919A1 true WO2007118919A1 (en) 2007-10-25

Family

ID=38609077

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/ES2007/000235 WO2007118919A1 (en) 2006-04-19 2007-04-19 Method for generating synthetic-animation images

Country Status (2)

Country Link
ES (1) ES2284391B1 (en)
WO (1) WO2007118919A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101916456A (en) * 2010-08-11 2010-12-15 李浩民 Method for producing personalized three-dimensional cartoon
CN101930618A (en) * 2010-08-20 2010-12-29 李浩民 Method for producing individual two-dimensional anime
CN102087750A (en) * 2010-06-13 2011-06-08 湖南宏梦信息科技有限公司 Method for manufacturing cartoon special effect

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0675461A2 (en) * 1994-03-22 1995-10-04 Casio Computer Co., Ltd. Method and apparatus for generating image
JPH0973559A (en) * 1995-09-07 1997-03-18 Fujitsu Ltd Morphing editing device
WO2003017206A1 (en) * 2001-08-14 2003-02-27 Pulse Entertainment, Inc. Automatic 3d modeling system and method
US20030051255A1 (en) * 1993-10-15 2003-03-13 Bulman Richard L. Object customization and presentation system

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030051255A1 (en) * 1993-10-15 2003-03-13 Bulman Richard L. Object customization and presentation system
EP0675461A2 (en) * 1994-03-22 1995-10-04 Casio Computer Co., Ltd. Method and apparatus for generating image
JPH0973559A (en) * 1995-09-07 1997-03-18 Fujitsu Ltd Morphing editing device
WO2003017206A1 (en) * 2001-08-14 2003-02-27 Pulse Entertainment, Inc. Automatic 3d modeling system and method

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102087750A (en) * 2010-06-13 2011-06-08 湖南宏梦信息科技有限公司 Method for manufacturing cartoon special effect
CN101916456A (en) * 2010-08-11 2010-12-15 李浩民 Method for producing personalized three-dimensional cartoon
CN101930618A (en) * 2010-08-20 2010-12-29 李浩民 Method for producing individual two-dimensional anime

Also Published As

Publication number Publication date
ES2284391A1 (en) 2007-11-01
ES2284391B1 (en) 2008-09-16

Similar Documents

Publication Publication Date Title
CN107154069B (en) Data processing method and system based on virtual roles
US10684467B2 (en) Image processing for head mounted display devices
CN109410298B (en) Virtual model manufacturing method and expression changing method
ES2237010T3 (en) PROCEDURE FOR THE CREATION OF 3D FACIAL MODELS FROM FACIAL IMAGES.
Magnenat-Thalmann et al. Handbook of virtual humans
CN107077755A (en) Virtually with real fusion method, system and virtual reality device
EP2580741A2 (en) Real-time animation of facial expressions
JP2004506276A (en) Three-dimensional face modeling system and modeling method
KR102215290B1 (en) Computer graphics synthesis system and method thereof
JP2012074878A (en) Image generation program, imaging device, imaging system, and image generation method
Gonzalez-Franco et al. Movebox: Democratizing mocap for the microsoft rocketbox avatar library
Hodgins et al. Computer animation
ES2284391B1 (en) PROCEDURE FOR THE GENERATION OF SYNTHETIC ANIMATION IMAGES.
JP2004171543A (en) Method and device for image processing
KR20140065762A (en) System for providing character video and method thereof
Aitken et al. The Lord of the Rings: the visual effects that brought middle earth to the screen
KR20200029968A (en) Modeling method of automatic character facial expression using deep learning technology
KR101780496B1 (en) Method for producing 3D digital actor image based on character modelling by computer graphic tool
CN110853147B (en) Three-dimensional face transformation method
Tiddeman et al. Transformation of dynamic facial image sequences using static 2D prototypes
Huang et al. A process for the semi-automated generation of life-sized, interactive 3D character models for holographic projection
Doroski Thoughts of Spirits in Madness: Virtual Production Animation and Digital Technologies for the Expansion of Independent Storytelling
JP2001319108A (en) Recreation, shopping, and business model
Morishima et al. Instant movie casting with personality: Dive into the movie system
US8896607B1 (en) Inverse kinematics for rigged deformable characters

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 07765824

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 07765824

Country of ref document: EP

Kind code of ref document: A1