US9262862B2 - Method and apparatus for reconstructing three dimensional model - Google Patents

Method and apparatus for reconstructing three dimensional model Download PDF

Info

Publication number
US9262862B2
US9262862B2 US13/686,927 US201213686927A US9262862B2 US 9262862 B2 US9262862 B2 US 9262862B2 US 201213686927 A US201213686927 A US 201213686927A US 9262862 B2 US9262862 B2 US 9262862B2
Authority
US
United States
Prior art keywords
substructures
model
local
dimensional model
depth images
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US13/686,927
Other versions
US20140099017A1 (en
Inventor
Yao-Yang Tsai
Hian-Kun Tenn
Jay Huang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Industrial Technology Research Institute ITRI
Original Assignee
Industrial Technology Research Institute ITRI
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Industrial Technology Research Institute ITRI filed Critical Industrial Technology Research Institute ITRI
Assigned to INDUSTRIAL TECHNOLOGY RESEARCH INSTITUTE reassignment INDUSTRIAL TECHNOLOGY RESEARCH INSTITUTE ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HUANG, JAY, TENN, HIAN-KUN, TSAI, YAO-YANG
Publication of US20140099017A1 publication Critical patent/US20140099017A1/en
Application granted granted Critical
Publication of US9262862B2 publication Critical patent/US9262862B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T7/0077
    • G06T7/0081
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/593Depth or shape recovery from multiple images from stereo images
    • G06T7/596Depth or shape recovery from multiple images from stereo images from three or more stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/08Indexing scheme for image data processing or generation, in general involving all processing steps from image acquisition to 3D model generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds

Definitions

  • the technical field relates to a method and an apparatus for reconstructing a three-dimensional model.
  • a three-dimensional (3D) model scanning technique is mainly used to obtain 3D information of an object, and reconstruct a 3D model on an electronic device capable of performing computations such as a computer or a handheld device, etc.
  • model edit software for example, Maya or 3DMax
  • not only an appearance of the 3D model has a higher fidelity, but also it has less fabrication time and low manpower demand.
  • the existing 3D model scanning technique mainly includes two core steps of “shooting” and “merging” images of an object.
  • a shooting angle of the object has to cover all possible angles as far as possible in order to guarantee integrity of a final result.
  • the “merging” step is executed to merge images captured in different angles into a 3D model.
  • the “merging” step is generally the most difficult step, and the largest difference in the existing techniques also lies in such step.
  • one of the existing techniques is to use a single camera to obtain shooting results of different time points, and calculate correlation of the shooting results according to a feature corresponding relationship of overlapped parts of the shooting results, so as to merge the shooting results to build a 3D model of the object.
  • another existing technique is to use a single camera and a turntable to record a rotating angle of a turntable corresponding to a shooting moment, and merge the shooting results of each angle obtained by the turntable, so to build the 3D model of the object.
  • another existing technique is to erect a plurality of cameras to cover all of the shooting angles, and simultaneously obtain shooting results of the object. Since positions of the cameras are all fixed, as long as a position and a shooting direction of each camera are obtained, shooting data of the cameras can be merged to build the 3D model of the object.
  • the single camera shoots the object at different time points, if an appearance of the object changes during the shooting period, the reconstructed model of the object is incomplete (for example, structure misalignment caused by change of the structure). Moreover, if a plurality of cameras are used to simultaneously shoot the object, the required cost is higher than the cost of using the single camera, which is difficult to be accepted by ordinary users.
  • the disclosure is directed to a method and an apparatus for reconstructing a three-dimensional (3D) model, by which a 3D model of a deformable object is capable of being accurately reconstructed.
  • An embodiment of the disclosure provides a method for reconstructing a three-dimensional (3D) model, which is adapted to build a 3D model of an object.
  • the method for reconstructing the 3D model includes following steps.
  • a plurality of first depth images of an object are obtained.
  • the first depth images of the object are divided into a plurality of depth image groups according to linking information of the object, where the linking information records location information corresponding to a plurality of substructures of the object, each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups.
  • a local model corresponding to each of the substructures is built according to the second depth images and the location information corresponding to each of the substructures.
  • the local models corresponding to the substructures are merged according to the linking information of the object, so as to build the 3D model of the object.
  • An embodiment of the disclosure provides a 3D model reconstructing apparatus, which is adapted to build a 3D model of an object.
  • the 3D model reconstructing apparatus includes an image capturing unit, a depth unit, an image grouping unit and a model building unit.
  • the image capturing unit captures shooting information of the object in different angles to provide a plurality of shooting results.
  • the depth unit is coupled to the image capturing unit, and builds a plurality of first depth images of the object according to the shooting results provided by the image capturing unit.
  • the image grouping unit is coupled to the depth unit, and divides the first depth images of the object into a plurality of depth image groups according to linking information of the object, where the linking information records location information corresponding to a plurality of substructures of the object, each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups.
  • the model building unit is coupled to the image grouping unit, and builds a local model corresponding to each of the substructures according to the second depth images and the location information corresponding to each of the substructures, and merges the local models corresponding to the substructures according to the linking information of the object, so as to build the 3D model of the object.
  • the 3D model reconstructing apparatus divides the first depth images of the object into a plurality of depth image groups according to the linking information of the object, where each of the first depth image groups has a plurality of second depth images. Moreover, the 3D model reconstructing apparatus builds a plurality of the local models of the object according to the depth image groups, and merges the local models to build the integral 3D model of the object. In this way, the 3D image of the deformable object is accurately built.
  • FIG. 1 is a block diagram of a three-dimensional (3D) model reconstructing apparatus system according to an exemplary embodiment of the disclosure.
  • FIGS. 2A-2C are schematic diagrams of shooting results of an object obtained by an image capturing unit according to an exemplary embodiment of the disclosure.
  • FIGS. 3A-3C are schematic diagrams of a plurality of first depth images of an object built by a depth unit according to an exemplary embodiment of the disclosure.
  • FIG. 4 is a schematic diagram of linking information of an object according to an exemplary embodiment of the disclosure.
  • FIGS. 5A-5B are schematic diagrams of a situation that an image grouping unit divides first depth images of an object into a plurality of depth image groups according to an exemplary embodiment of the disclosure.
  • FIG. 5C is a schematic diagram illustrating a situation that a model building unit builds a local model corresponding to each substructure according to an exemplary embodiment of the disclosure.
  • FIG. 5D is a schematic diagram illustrating a situation that a model building unit merges local models to build a 3D model of an object according to an exemplary embodiment of the disclosure.
  • FIGS. 5E-5G are schematic diagrams of local models of an object built by a model building unit viewed from different viewing-angles according to an exemplary embodiment of the disclosure.
  • FIG. 5H is a schematic diagram of a connection relationship of a 3D model built by a model building unit according to an exemplary embodiment of the disclosure.
  • FIG. 6 is a flowchart illustrating a method for reconstructing a 3D model according to an exemplary embodiment of the disclosure.
  • FIG. 1 is a block diagram of a three-dimensional (3D) model reconstructing apparatus system according to an exemplary embodiment of the disclosure.
  • the 3D model reconstructing apparatus system 1000 includes a 3D model reconstructing apparatus 10 and an object 200 , where the 3D model reconstructing apparatus 10 is used to build a 3D model of the object 200 .
  • the 3D model reconstructing apparatus 10 is, for example, an electronic apparatus such as a notebook computer, a tablet PC, a personal digital assistant (PDA), a mobile phone, a digital camera, and e-book, or a game machine, etc., which is not limited by the disclosure.
  • the 3D model reconstructing apparatus 10 includes a processor 102 , a memory 104 , an image capturing unit 110 , a depth unit 120 , and image grouping unit 130 and a model building unit 140 .
  • the 3D model reconstructing apparatus 10 further includes a structure recognizing unit 135 , a coordinate transforming unit 150 and a display unit 160 . Functions of the above components are respectively described below.
  • the processor 102 can be hardware (for example, a chipset or a processor, etc.) having computation capability, which is used to control a whole operation of the 3D model reconstructing apparatus 10 .
  • the processor 102 is, for example, a central processing unit (CPU), or other programmable device, for example, a microprocessor, a digital signal processor (DSP), a programmable controller, an application specific integrated circuit (ASIC), a programmable logic device (PLD) or other similar devices.
  • CPU central processing unit
  • DSP digital signal processor
  • ASIC application specific integrated circuit
  • PLD programmable logic device
  • the memory 104 is coupled to the processor 102 .
  • the memory 104 can be an internal storage unit or an external storage unit.
  • the internal storage unit can be a random access memory (RAM), a read-only memory (ROM), a flash memory, a magnetic disk storage device, etc.
  • the external storage unit can be a compact flash (CF), a memory card, a secure digital (SD) memory card, a micro SD memory card, or a memory stick (MS), etc., though the disclosure is not limited thereto.
  • the memory 104 can be used to store image information of the object 200 or information required for building the 3D model of the object 200 .
  • the image capturing unit 110 can be a video camera or a camera using a charge coupled device (CCD) lens, a complementary metal oxide semiconductor transistor (CMOS) lens or an infrared lens, though the disclosure is not limited thereto.
  • the image capturing unit 110 is used to capture shooting information of the object 200 in different angles to provide a plurality of shooting results. Namely, regarding an object 200 capable of changing an action pattern thereof, within a shooting time that the image capturing unit 110 shoots the object 200 , the image capturing unit 110 can capture different shooting results presented by the object 200 along a fixed or unfixed shooting direction when the object 200 conducts a dynamic motion to generate a motion pattern variation.
  • the shooting direction of the image capturing unit 110 of the present exemplary embodiment is fixed or unfixed, when the object 200 conducts a dynamic motion, an angle between the object 200 and the image capturing unit 110 is varied along with the shooting time, such that the image capturing unit 110 can capture images of the object 200 that faces the image capturing unit 110 in different angles.
  • the aforementioned shooting time is, for example, 30 seconds, and the image capturing unit 110 , for example, captures an image of the object 200 every 0.03 seconds within the shooting time, namely, the image capturing unit 110 can obtain 1000 pieces of shooting information of the object 200 shot in different angles within the shooting time, though the disclosure is not limited thereto.
  • FIGS. 2A-2C are schematic diagrams of shooting results of the object 200 obtained by the image capturing unit 110 according to an exemplary embodiment of the disclosure. It is assumed that the object 200 shot by the image capturing unit 110 is a human body conducting a dynamic motion, and the image capturing unit 110 obtains three shooting results within the shooting time.
  • the image capturing unit 110 obtains a plurality of shooting results 20 a , 20 b and 20 c of the object 200 in different angles.
  • the shooting result 20 a shown in FIG. 2A corresponds to a profile 202 of the object 200
  • the shooting result 20 b shown in FIG. 2B corresponds to another profile 204 of the object 200
  • the shooting result 20 c shown in FIG. 2C corresponding to a front side 202 of the object 200 , where a y-direction represents a direction that the front side 206 of the object 200 faces to.
  • the depth unit 120 is coupled to the image capturing unit 110 .
  • the depth unit 120 builds a plurality of first depth images of the object 200 according to the shooting results provided by the image capturing unit 110 , where the first depth image has depth information of the object 200 .
  • each of the first depth images has a plurality of pixels, where each pixel can be composed of three values (for example, floating-point numbers), which respectively represent a distance between a surface of the object 200 of each pixel and the image capturing unit 110 .
  • each of the first depth images can present position relationships between the object 200 and the image capturing unit 110 through the pixels, i.e. the so-called depth information of the object 200 .
  • FIGS. 3A-3C are schematic diagrams of a plurality of first depth images of the object 200 built by the depth unit 120 according to an exemplary embodiment of the disclosure, where the depth unit 120 builds first depth images 30 a , 30 b and 30 c of the object 200 according to the shooting results 20 a , 20 b and 20 c provided by the image capturing unit 110 .
  • the first depth image 30 a of the object 200 built by the depth unit 120 corresponds the shooting result 20 a of the profile 202 of the object 200 .
  • FIG. 3A the first depth image 30 a of the object 200 built by the depth unit 120 corresponds the shooting result 20 a of the profile 202 of the object 200 .
  • the first depth image 30 b of the object 200 built by the depth unit 120 corresponds the shooting result 20 b of the profile 204 of the object 200 .
  • the first depth image 30 c of the object 200 built by the depth unit 120 corresponds the shooting result 20 c of the front side 206 of the object 200 .
  • the image grouping unit 130 is coupled to the depth unit 120 .
  • the image grouping unit 130 divides the first depth images of the object 200 into a plurality of depth image groups according to linking information of the object 200 , where the linking information records location information corresponding to a plurality of substructures of the object 200 , each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups.
  • the object 200 of the present exemplary embodiment may have a plurality of substructures, where each of the substructures is, for example, a local block of the object 200 .
  • each of the substructures may have a plurality of microstructures, where each of the macrostructures can be regarded as a structure unit of the object 200 that cannot be further divided. Therefore, the linking information of the object 200 records a connection relationship of the substructures of the object 200 and records location information of the microstructures of each of the substructures.
  • the object 200 of the present exemplary embodiment may have a global coordinate system, and the location information of each of the substructures on the global coordinate system is referred to as first coordinate information of the substructure, namely, the first coordinate information of each of the substructures is obtained according to the global coordinate system where the object 200 locates. Moreover, the location information of the microstructures of each of the substructures is also on such total coordinate system. In this way, since each of the substructures has the first coordinate information on the global coordinate system, the image grouping unit 130 can learn the connection relationship between the substructures according to the linking information, and a location of each of the substructures on the object 200 and location information thereof.
  • FIG. 4 is a schematic diagram of the linking information of the object 200 according to an exemplary embodiment of the disclosure, where a situation that the object 200 is a human body is taken as an example to describe the linking information of the present exemplary embodiment.
  • the object 200 has substructures p 1 ⁇ p 11 , where the substructures p 1 -p 11 have at least one or a plurality of microstructures p 1 - 1 ⁇ p 1 -N 11 , p 2 - 1 ⁇ p 2 -N 2 , . . . , p 11 - 1 ⁇ p 11 -N 11 , where N 1 ⁇ N 11 are respectively positive integers.
  • each of the substructures p 1 ⁇ p 11 can be regarded as a rigid body, namely, during a process that the image capturing unit 110 shoots the object 200 , although the object 200 can conduct the dynamic motion, the microstructures p 1 - 1 ⁇ p 1 -N 1 , p 2 - 1 ⁇ p 2 -N 2 , . . . , p 11 - 1 ⁇ p 11 -N 11 in the substructures p 1 ⁇ p 11 may respectively have a same moving speed, namely, each of the substructures p 1 ⁇ p 11 is regarded to have a single speed.
  • the object 200 is, for example, the human body
  • each of the substructures p 1 -p 11 is, for example, a local skeleton structure of the human body
  • the microstructures p 1 - 1 ⁇ p 1 -N 1 , p 2 - 1 ⁇ p 2 -N 2 , . . . , p 11 - 1 ⁇ p 11 -N 11 in the substructures p 1 ⁇ p 11 corresponds to each bone of the human body (i.e. the aforementioned structure unit).
  • each of the substructures p 1 ⁇ p 11 is, for example, a skeleton structure of a head of the human body, a skeleton structure of a front arm or upper arm, a skeleton structure of a chest, a skeleton structure of an abdomen, or a skeleton structure of a thigh or lower leg, though the disclosure is not limited thereto.
  • the substructure p 3 is for example, the front arm, and the microstructures p 3 - 1 ⁇ p 3 -N 3 of the substructure p 3 are, for example, a radius bone, an ulna and a finger bone, etc.
  • each of the substructures p 1 ⁇ p 11 of the object 200 can be regarded as a local skeleton structure having a plurality of bones. Therefore, in the present exemplary embodiment, the linking information records location information of a part of the microstructures (for example, bones) in the object 200 , so as to define the microstructures belonging to one substructure (for example, the local skeleton structure), and according to the local information of each of the microstructures, the linking information may also include the connection relationship between the substructures and the location information of each of the substructures in the object 200 (i.e. the global coordinate system of the object 200 ). In this way, the object 200 can be divided into the substructures p 1 ⁇ p 11 according to the linking information.
  • the structure recognizing unit 135 can be used to produce the substructures p 1 ⁇ p 11 of the object 200 , where the structure recognizing unit 135 can be respectively coupled to the image grouping unit 130 and the depth unit 120 . Further, the structure recognizing unit 135 can recognize positions of each of the microstructures p 1 - 1 ⁇ p 1 -N 1 , . . .
  • the structure recognizing unit 135 can estimate the positions of the pixels corresponding to the microstructures p 1 - 1 ⁇ p 1 -N 1 , . . .
  • the structure recognizing unit 135 can divide the substructures p 1 ⁇ p 11 of the object 200 by calculating the corresponding relationship between the pixels and the microstructures p 1 - 1 ⁇ p 1 -N 1 , . . . , p 11 - 1 ⁇ p 11 -N 11 in the first depth images 30 a , 30 b and 30 c .
  • the structure recognizing unit 135 can detect and track the appearance of the object 200 with a variable motion pattern by taking the microstructures p 1 - 1 ⁇ p 1 -N 1 , . . . , p 11 - 1 ⁇ p 11 -N 11 as a sample and using the connection relationship between the substructures p 1 ⁇ p 11 in the object 200 .
  • the object 200 is the human body
  • the structure recognizing unit 135 recognizes that the pixel is located at an extremity of the four limbs.
  • the structure recognizing unit 135 recognizes that the pixel is located at a torso portion of the object 200 .
  • the structure recognizing unit 135 recognizes that the pixel is located at a bone area of the object 200 . In this way, according to the first depth images 30 a , 30 b and 30 c of the object 200 and the location information of the microstructures p 1 - 1 ⁇ p 1 -N 1 , . . .
  • the structure recognizing unit 135 can use the first depth images 30 a , 30 b and 30 to recognize the microstructures p 1 - 1 ⁇ p 1 -N 1 , . . . , p 11 - 1 ⁇ p 11 -N 11 of the object 200 , and track the appearance of the object 200 . Therefore, the structure recognizing unit 135 can recognize the microstructures that have the same speed during the process that the image capturing unit 110 shoots the object 200 , so as to find the substructures composed of the microstructures.
  • the image grouping unit 130 divides each of the first depth images 30 a , 30 b and 30 c of the object 200 into a plurality of second depth images according to the substructures p 1 ⁇ p 11 of the object 200 and the depth values of the first depth images 30 a , 30 b and 30 c , and takes the corresponding second depth images in the first depth images 30 a , 30 b and 30 c as an image group, so as to divide depth image groups located at local blocks of each of the substructures p 1 ⁇ p 11 of the object 200 .
  • FIGS. 5A-5B are schematic diagrams of a situation that the image grouping unit 130 divides the first depth images of the object 200 into a plurality of depth image groups according to an exemplary embodiment of the disclosure, which are described below with reference of the first depth images 30 a , 30 b and 30 c of FIGS. 3A-3B and the substructures of FIG. 4 .
  • the image grouping unit 130 divides each of the first depth images 30 a , 30 b and 30 c of the object 200 into a plurality of second depth images 501 ⁇ 511 according to the divided substructures p 1 ⁇ p 11 of the object 200 and the depth values of the first depth images 30 a , 30 b and 30 c , and takes the corresponding second depth images 501 ⁇ 511 in the first depth images 30 a , 30 b and 30 c as an image group, so as to divide a plurality of depth image groups 501 s ⁇ 511 s.
  • the image grouping unit 130 divides the first depth image 30 a into the second depth images 501 ⁇ 511 according to the substructures p 1 ⁇ p 11 of the linking information of the object 200 , for example, depth images corresponding to the head, the front arm or upper arm, the chest, the abdomen, the thigh or lower leg of the human body.
  • the image grouping unit 130 divides first depth images 30 b and 30 c into the second depth images 501 ⁇ 511 the according to the substructures p 1 ⁇ p 11 of the linking information of the object 200 .
  • taking the second depth image 501 i.e.
  • the image grouping unit 130 takes the second depth images 501 of the first depth images 30 a , 30 b and 30 c as a depth image group, and the other second depth images 502 ⁇ 511 can be deduced by analogy. In this way, the image grouping unit 130 generates the depth image groups 501 s ⁇ 511 s on the substructures p 1 ⁇ p 11 .
  • the model building unit 140 is coupled to the image grouping unit 130 .
  • the model building unit 140 builds a local model corresponding to each of the substructures p 1 ⁇ p 11 on the object 200 according to the second depth images 501 ⁇ 511 and the location information corresponding to each of the substructures p 1 ⁇ p 11 .
  • the model building unit 140 merges the local models corresponding to the substructures p 1 ⁇ p 11 according to the linking information of the object 200 , so as to build the 3D model of the object 200 .
  • FIG. 5C is a schematic diagram illustrating a situation that the model building unit 140 builds a local model corresponding to each of the substructures according to an exemplary embodiment of the disclosure.
  • FIG. 5D is a schematic diagram illustrating a situation that the model building unit 140 merges the local models to build the 3D model of the object 200 according to an exemplary embodiment of the disclosure.
  • the model building unit 140 obtains location information and orientation information of the microstructures in each of the substructures p 1 ⁇ p 11 relative to the local coordinate systems C 1 ⁇ C 11 , and respectively generates local models M 1 ⁇ M 11 of the object 200 according to the depth image groups 501 s ⁇ 511 s corresponding to the substructures p 1 ⁇ p 11 .
  • the object 200 is the human body
  • the local model M 1 corresponds to the head
  • the local models M 2 ⁇ M 3 , M 5 ⁇ M 6 correspond to the front arm and the upper arms
  • the local model M 4 corresponds to the chest
  • the local model M 7 corresponds to the abdomen
  • the local models M 8 ⁇ M 11 correspond to thighs and the lower legs.
  • the 3D model reconstructing apparatus 10 of the present exemplary embodiment further selectively includes the coordinate transforming unit 150 .
  • the coordinate transforming unit 150 is coupled to the image grouping unit 130 and the model building unit 140 .
  • the coordinate transforming unit 150 transforms the first coordinate information of each of the substructures from the global coordinate system Co of the object 200 to the local coordinate systems C 1 ⁇ C 11 (for example, cylindrical coordinate systems) of the substructures p 1 ⁇ p 11 through a transforming calculation, where the local coordinate systems C 1 ⁇ C 11 correspond to the substructures p 1 ⁇ p 11 , and the substructures p 1 ⁇ p 11 have the second coordinate information in the local coordinate systems C 1 ⁇ C 11 .
  • the substructures p 1 ⁇ p 11 of the object 200 respectively have their own local coordinate systems C 1 ⁇ C 11 , and each of the substructures p 1 ⁇ p 11 have location information of the local coordinate systems C 1 ⁇ C 11 , which is referred to as second coordinate information.
  • each of the substructures and the microstructures in each of the substructures can be regarded as a rigid body, namely, the appearance of each of the microstructures is not changed (for example, the appearance of the of the microstructures is not elongated, shortened or distorted) during the shooting process of the image capturing unit 110 .
  • the 3D coordinate information of the pixels in the depth image can be stored according to the respective local coordinate system of each of the substructures of the object 200 .
  • the model building unit 140 can merge the depth image groups 501 s ⁇ 511 s corresponding to the substructures p 1 ⁇ p 11 according to the second coordinate information of the local coordinate systems C 1 ⁇ C 11 , so as to build the local models M 1 ⁇ M 11 corresponding to each of the substructures p 1 ⁇ p 11 .
  • the model building unit 140 when the model building unit 140 builds the local models M 1 ⁇ M 11 corresponding to each of the substructures p 1 ⁇ p 11 , the model building unit 140 further determines whether each of the second depth images 501 ⁇ 511 in the depth image groups 501 s ⁇ 511 s have an overlapped region. If the second depth images 501 ⁇ 511 have the overlapped region, the model building unit 140 excludes the overlapped region, and builds the local models according to remained depth images in the second depth images 501 - 511 excluding the overlapped region. For example, regarding the depth image group 501 s of FIG.
  • the model building unit 140 excludes the overlapped region, and uses the remained depth images of the second depth images 501 excluding the overlapped region to build the local models M 1 ⁇ M 11 of the object 200 in the substructure p 1 .
  • each of the second depth images 501 ⁇ 511 in the depth image groups 501 s ⁇ 511 s have an overlapped region
  • the model building unit 140 merges the second depth images 501 ⁇ 511 in each of the depth image groups
  • the model building unit 140 adjusts a configuration of each of the second depth images 501 ⁇ 511 of the depth image groups 501 s ⁇ 511 s on each of the local coordinate systems according to feature information of the object 200 on the surface of each of the substructures p 1 ⁇ p 11 , so as to build the local models M 1 ⁇ M 11 corresponding to the substructures p 1 ⁇ p 11 .
  • the feature information is, for example, a surface texture of the object 200 , or a profile of the object 200 , etc., which is not limited by the disclosure.
  • the model building unit 140 further finely adjusts the partial models M 1 ⁇ M 11 according to the features of the surface texture or the profile, etc. of the object 200 . In this way, the model building unit 140 can minimize the overlapped region between the second depth images in each of the depth image groups, so as to build the accurate local models M 1 ⁇ M 11 .
  • FIGS. 5E-5G are schematic diagrams of local models of the object 200 built by the model building unit 140 viewed from different viewing-angles according to an exemplary embodiment of the disclosure.
  • the local model M 4 of the object 200 in FIG. 5D is taken as an example for descriptions, and the local model M 4 is a chest model of the object 200 .
  • the local model M 4 is generated by the model building unit 140 according to the depth image group 504 s . As shown in FIGS.
  • the local model M 4 can be composed of a second depth image 504 a (coming from the first depth image 30 a ), a second depth image 504 b (coming from the first depth image 30 b ), and a second depth image 504 c (coming from the first depth image 30 c ).
  • the model building unit 140 further merges the local models M 1 ⁇ M 11 on the local coordinate systems of the substructures according to the linking information of the object 200 to transform the local models M 1 ⁇ M 11 of the substructures from the local coordinate systems to the global coordinate system Co of the object 200 , so as to merge the local models M 1 ⁇ M 11 corresponding to the substructures to build the 3D model of the object 200 .
  • the model building unit 140 transforms the depth image groups corresponding to the substructures from the local coordinate systems to the global coordinate system Co of the object 200 according to the second coordinate information of the substructures on the local coordinate systems, and merges the local models M 1 ⁇ M 11 to generate a 3D model Mf of the integral object 200 .
  • the model building unit 140 further merges the local models M 1 ⁇ M 11 according to the connection relationship between the microstructures of each of the substructures.
  • the model building unit 140 takes each pixel in the second depth image corresponding to each of the substructures as a point, and during a process that the model building unit 140 merges the local models M 1 ⁇ M 11 , and when three neighboring points are connected to another point, the model building unit 140 determines whether a first figure formed by the three neighboring points is overlapped with a second figure formed by two of the three neighboring points and the other point.
  • the model building unit 140 takes a triangle formed by each point and two neighboring points as a unit for connection (i.e. form a plurality of triangles), and determines whether an overlapping phenomenon occurs between the connected triangles, i.e. checks whether the 3D planes represented by the triangles have spatial relationships of plane intersection and plane overlapping. If the 3D planes represented by the triangles have spatial relationships of plane intersection and plane overlapping, such connection relationship is incorrect, and the model building unit 140 selects another point to execute the above determination and check operations. If the overlapping phenomenon is not occurred, the model building unit 140 determines that the connection relationship between the checked three points is correct.
  • FIG. 5H is a schematic diagram of the connection relationship of the 3D model Mf built by the model building unit 140 according to an exemplary embodiment of the disclosure.
  • the model building unit 140 determines whether the triangles formed by each three points in the points Q 1 -Q 4 have the overlapping phenomenon. For example, regarding a triangle T 1 formed by the point Q 1 and the neighboring points Q 3 and Q 4 , and a triangle T 2 formed by the point Q 2 and the neighboring points Q 3 and Q 4 , since the overlapping phenomenon occurs between the triangle T 1 and the triangle T 2 , the model building unit 140 determines that the connection relationship of the triangle T 1 and the triangle T 2 (i.e. the connection relationship of the points Q 1 , Q 3 and Q 4 and the connection relationship of the points Q 2 , Q 3 and Q 4 ) is incorrect.
  • the connection relationship of the triangle T 1 and the triangle T 2 ′ i.e. the connection relationship of the points Q 1 , Q 3 and Q 4 and the connection relationship of the points Q 2 , Q 1 and Q 4 .
  • the model building unit 140 checks whether a triangle formed by each three points is overlapped with the neighboring triangles until all of the points and the neighboring points have the connection relationships.
  • the display unit 160 can be selectively configured for coupling to the model building unit 140 .
  • the display unit 160 can display the 3D model of the object 200 built by the model building unit 140 . In this way, a user can view the accurate 3D model of the object 200 through the display unit 160 .
  • the 3D model reconstructing apparatus 10 of the present exemplary embodiment obtains shooting information of the object 200 in different angles through the image capturing unit 110 , and generates a plurality of the first depth images 30 a - 30 c of the object 200 through the depth unit 120 . Moreover, the 3D model reconstructing apparatus 10 divides the first depth images 30 a - 30 c into a plurality of depth image groups 501 s ⁇ 511 s according to the linking information of the object 200 . Particularly, the 3D model reconstructing apparatus 10 generates the local models M 1 ⁇ M 11 according to the coordinate systems of the depth image groups 501 s ⁇ 511 s .
  • the 3D model reconstructing apparatus 10 merges the local models M 1 ⁇ M 11 to generate the 3D model Mf of the object 200 .
  • an accurate 3D model of a deformable object can be built through the image capturing unit 110 .
  • FIG. 6 is a flowchart illustrating a method for reconstructing a 3D model according to an exemplary embodiment of the disclosure.
  • step S 602 the image capturing unit 110 is used to capture shooting information of the object 200 in different angles to provide a plurality of shooting results.
  • step S 604 the depth unit 120 builds a plurality of first depth images of the object 200 according to the shooting results provided by the image capturing unit 110 , where the first depth image has depth information of the object 200 .
  • step S 606 the structure recognizing unit 135 calculates a corresponding relationship between a plurality of microstructures in the object 200 and the first depth images according to linking information of the object 200 , so as to divide a plurality of substructures of the object 200 .
  • step S 608 the image grouping unit 130 divides the first depth images of the object 200 into a plurality of depth image groups according to the linking information and the substructures of the object 200 , where the linking information records location information corresponding to a plurality of substructures of the object 200 , each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups.
  • step S 610 the model building unit 140 determines whether an overlapping region exists between the second depth images of each of the depth image groups.
  • a step S 612 is executed, by which the model building unit 140 builds a local model corresponding to each of the substructures according to the second depth images and the location information corresponding to each of the substructures.
  • a step S 614 is executed, by which the model building unit 140 excludes the overlapping region, and builds the local model according to remained depth images in the second depth images excluding the overlapped region by using feature information of the object 200 on the surface of each of the substructures.
  • step S 616 the model building unit 140 transforms the local models of the substructures from the local coordinate systems to the global coordinate system of the object 200 .
  • step S 618 the model building unit 140 merges the local models on the local coordinate systems of the substructures according to the linking information of the object 200 , so as to build the 3D model of the object 200 .
  • the first depth images of the object in different angles are divided into a plurality of depth image groups according to the linking information of the object, where each of the first depth image groups has a plurality of second depth images.
  • each of the depth image groups is transformed from the global coordinate system of the object to the respective local coordinate systems of the depth image groups, so as to generate the local model by using each of the depth image groups.
  • an integral 3D model of the object 200 is built by merging the local models. In this way, the 3D image of the deformable object is accurately built.

Abstract

A method and an apparatus for reconstructing a three dimensional model of an object are provided. The method includes the following steps. A plurality of first depth images of an object are obtained. According to a linking information of the object, the first depth images are divided into a plurality of depth image groups. The linking information records location information corresponding to a plurality of substructures of the object. Each depth image group includes a plurality of second depth images, and the substructures correspond to the second depth images. According to the second depth image and the location information corresponding to each substructure, a local module of each substructure is built. According to the linking information, the local models corresponding to the substructures are merged, and the three-dimensional model of the object is built.

Description

CROSS-REFERENCE TO RELATED APPLICATION
This application claims the priority benefit of Taiwan application serial no. 101136721, filed on Oct. 4, 2012. The entirety of the above-mentioned patent application is hereby incorporated by reference herein and made a part of this specification.
TECHNICAL FIELD
The technical field relates to a method and an apparatus for reconstructing a three-dimensional model.
BACKGROUND
A three-dimensional (3D) model scanning technique is mainly used to obtain 3D information of an object, and reconstruct a 3D model on an electronic device capable of performing computations such as a computer or a handheld device, etc. Compared to a model produced by an art staff by using model edit software (for example, Maya or 3DMax), not only an appearance of the 3D model has a higher fidelity, but also it has less fabrication time and low manpower demand.
The existing 3D model scanning technique mainly includes two core steps of “shooting” and “merging” images of an object. For example, in the “shooting” step, a shooting angle of the object has to cover all possible angles as far as possible in order to guarantee integrity of a final result. After the “shooting” step is completed, the “merging” step is executed to merge images captured in different angles into a 3D model. In the two core steps, the “merging” step is generally the most difficult step, and the largest difference in the existing techniques also lies in such step.
For example, one of the existing techniques is to use a single camera to obtain shooting results of different time points, and calculate correlation of the shooting results according to a feature corresponding relationship of overlapped parts of the shooting results, so as to merge the shooting results to build a 3D model of the object. Alternatively, another existing technique is to use a single camera and a turntable to record a rotating angle of a turntable corresponding to a shooting moment, and merge the shooting results of each angle obtained by the turntable, so to build the 3D model of the object. Moreover, another existing technique is to erect a plurality of cameras to cover all of the shooting angles, and simultaneously obtain shooting results of the object. Since positions of the cameras are all fixed, as long as a position and a shooting direction of each camera are obtained, shooting data of the cameras can be merged to build the 3D model of the object.
However, in the above existing techniques, since the single camera shoots the object at different time points, if an appearance of the object changes during the shooting period, the reconstructed model of the object is incomplete (for example, structure misalignment caused by change of the structure). Moreover, if a plurality of cameras are used to simultaneously shoot the object, the required cost is higher than the cost of using the single camera, which is difficult to be accepted by ordinary users.
Therefore, it is an important problem to be resolved by manufactures to accurately reconstruct a 3D model of a deformable object.
SUMMARY
The disclosure is directed to a method and an apparatus for reconstructing a three-dimensional (3D) model, by which a 3D model of a deformable object is capable of being accurately reconstructed.
An embodiment of the disclosure provides a method for reconstructing a three-dimensional (3D) model, which is adapted to build a 3D model of an object. The method for reconstructing the 3D model includes following steps. A plurality of first depth images of an object are obtained. The first depth images of the object are divided into a plurality of depth image groups according to linking information of the object, where the linking information records location information corresponding to a plurality of substructures of the object, each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups. A local model corresponding to each of the substructures is built according to the second depth images and the location information corresponding to each of the substructures. The local models corresponding to the substructures are merged according to the linking information of the object, so as to build the 3D model of the object.
An embodiment of the disclosure provides a 3D model reconstructing apparatus, which is adapted to build a 3D model of an object. The 3D model reconstructing apparatus includes an image capturing unit, a depth unit, an image grouping unit and a model building unit. The image capturing unit captures shooting information of the object in different angles to provide a plurality of shooting results. The depth unit is coupled to the image capturing unit, and builds a plurality of first depth images of the object according to the shooting results provided by the image capturing unit. The image grouping unit is coupled to the depth unit, and divides the first depth images of the object into a plurality of depth image groups according to linking information of the object, where the linking information records location information corresponding to a plurality of substructures of the object, each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups. The model building unit is coupled to the image grouping unit, and builds a local model corresponding to each of the substructures according to the second depth images and the location information corresponding to each of the substructures, and merges the local models corresponding to the substructures according to the linking information of the object, so as to build the 3D model of the object.
According to the above descriptions, the 3D model reconstructing apparatus divides the first depth images of the object into a plurality of depth image groups according to the linking information of the object, where each of the first depth image groups has a plurality of second depth images. Moreover, the 3D model reconstructing apparatus builds a plurality of the local models of the object according to the depth image groups, and merges the local models to build the integral 3D model of the object. In this way, the 3D image of the deformable object is accurately built.
In order to make the aforementioned and other features and advantages of the disclosure comprehensible, several exemplary embodiments accompanied with figures are described in detail below.
BRIEF DESCRIPTION OF THE DRAWINGS
The accompanying drawings are included to provide a further understanding of the disclosure, and are incorporated in and constitute a part of this specification. The drawings illustrate embodiments of the disclosure and, together with the description, serve to explain the principles of the disclosure.
FIG. 1 is a block diagram of a three-dimensional (3D) model reconstructing apparatus system according to an exemplary embodiment of the disclosure.
FIGS. 2A-2C are schematic diagrams of shooting results of an object obtained by an image capturing unit according to an exemplary embodiment of the disclosure.
FIGS. 3A-3C are schematic diagrams of a plurality of first depth images of an object built by a depth unit according to an exemplary embodiment of the disclosure.
FIG. 4 is a schematic diagram of linking information of an object according to an exemplary embodiment of the disclosure.
FIGS. 5A-5B are schematic diagrams of a situation that an image grouping unit divides first depth images of an object into a plurality of depth image groups according to an exemplary embodiment of the disclosure.
FIG. 5C is a schematic diagram illustrating a situation that a model building unit builds a local model corresponding to each substructure according to an exemplary embodiment of the disclosure.
FIG. 5D is a schematic diagram illustrating a situation that a model building unit merges local models to build a 3D model of an object according to an exemplary embodiment of the disclosure.
FIGS. 5E-5G are schematic diagrams of local models of an object built by a model building unit viewed from different viewing-angles according to an exemplary embodiment of the disclosure.
FIG. 5H is a schematic diagram of a connection relationship of a 3D model built by a model building unit according to an exemplary embodiment of the disclosure.
FIG. 6 is a flowchart illustrating a method for reconstructing a 3D model according to an exemplary embodiment of the disclosure.
DETAILED DESCRIPTION OF DISCLOSED EMBODIMENTS
In the following detailed description, for purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the disclosed embodiments. It will be apparent, however, that one or more embodiments may be practiced without these specific details. In other instances, well-known structures and devices are schematically shown in order to simplify the drawing.
FIG. 1 is a block diagram of a three-dimensional (3D) model reconstructing apparatus system according to an exemplary embodiment of the disclosure.
Referring to FIG. 1, the 3D model reconstructing apparatus system 1000 includes a 3D model reconstructing apparatus 10 and an object 200, where the 3D model reconstructing apparatus 10 is used to build a 3D model of the object 200. The 3D model reconstructing apparatus 10 is, for example, an electronic apparatus such as a notebook computer, a tablet PC, a personal digital assistant (PDA), a mobile phone, a digital camera, and e-book, or a game machine, etc., which is not limited by the disclosure. In the present exemplary embodiment, the 3D model reconstructing apparatus 10 includes a processor 102, a memory 104, an image capturing unit 110, a depth unit 120, and image grouping unit 130 and a model building unit 140. Moreover, the 3D model reconstructing apparatus 10 further includes a structure recognizing unit 135, a coordinate transforming unit 150 and a display unit 160. Functions of the above components are respectively described below.
The processor 102 can be hardware (for example, a chipset or a processor, etc.) having computation capability, which is used to control a whole operation of the 3D model reconstructing apparatus 10. In the present exemplary embodiment, the processor 102 is, for example, a central processing unit (CPU), or other programmable device, for example, a microprocessor, a digital signal processor (DSP), a programmable controller, an application specific integrated circuit (ASIC), a programmable logic device (PLD) or other similar devices.
The memory 104 is coupled to the processor 102. The memory 104 can be an internal storage unit or an external storage unit. The internal storage unit can be a random access memory (RAM), a read-only memory (ROM), a flash memory, a magnetic disk storage device, etc. The external storage unit can be a compact flash (CF), a memory card, a secure digital (SD) memory card, a micro SD memory card, or a memory stick (MS), etc., though the disclosure is not limited thereto. In the present exemplary embodiment, the memory 104 can be used to store image information of the object 200 or information required for building the 3D model of the object 200.
The image capturing unit 110 can be a video camera or a camera using a charge coupled device (CCD) lens, a complementary metal oxide semiconductor transistor (CMOS) lens or an infrared lens, though the disclosure is not limited thereto. The image capturing unit 110 is used to capture shooting information of the object 200 in different angles to provide a plurality of shooting results. Namely, regarding an object 200 capable of changing an action pattern thereof, within a shooting time that the image capturing unit 110 shoots the object 200, the image capturing unit 110 can capture different shooting results presented by the object 200 along a fixed or unfixed shooting direction when the object 200 conducts a dynamic motion to generate a motion pattern variation.
In detail, since the shooting direction of the image capturing unit 110 of the present exemplary embodiment is fixed or unfixed, when the object 200 conducts a dynamic motion, an angle between the object 200 and the image capturing unit 110 is varied along with the shooting time, such that the image capturing unit 110 can capture images of the object 200 that faces the image capturing unit 110 in different angles. The aforementioned shooting time is, for example, 30 seconds, and the image capturing unit 110, for example, captures an image of the object 200 every 0.03 seconds within the shooting time, namely, the image capturing unit 110 can obtain 1000 pieces of shooting information of the object 200 shot in different angles within the shooting time, though the disclosure is not limited thereto.
FIGS. 2A-2C are schematic diagrams of shooting results of the object 200 obtained by the image capturing unit 110 according to an exemplary embodiment of the disclosure. It is assumed that the object 200 shot by the image capturing unit 110 is a human body conducting a dynamic motion, and the image capturing unit 110 obtains three shooting results within the shooting time.
Referring to FIGS. 2A-2C, along with motion of the object 200, the image capturing unit 110 obtains a plurality of shooting results 20 a, 20 b and 20 c of the object 200 in different angles. For example, the shooting result 20 a shown in FIG. 2A corresponds to a profile 202 of the object 200, the shooting result 20 b shown in FIG. 2B corresponds to another profile 204 of the object 200, and the shooting result 20 c shown in FIG. 2C corresponding to a front side 202 of the object 200, where a y-direction represents a direction that the front side 206 of the object 200 faces to.
The depth unit 120 is coupled to the image capturing unit 110. The depth unit 120 builds a plurality of first depth images of the object 200 according to the shooting results provided by the image capturing unit 110, where the first depth image has depth information of the object 200. In detail, each of the first depth images has a plurality of pixels, where each pixel can be composed of three values (for example, floating-point numbers), which respectively represent a distance between a surface of the object 200 of each pixel and the image capturing unit 110. In other words, each of the first depth images can present position relationships between the object 200 and the image capturing unit 110 through the pixels, i.e. the so-called depth information of the object 200. Moreover, when each of the pixels in the first depth image is composed of three floating-point numbers, the first depth image may have the depth information of the object 200 with higher resolution. FIGS. 3A-3C are schematic diagrams of a plurality of first depth images of the object 200 built by the depth unit 120 according to an exemplary embodiment of the disclosure, where the depth unit 120 builds first depth images 30 a, 30 b and 30 c of the object 200 according to the shooting results 20 a, 20 b and 20 c provided by the image capturing unit 110. As shown in FIG. 3A, the first depth image 30 a of the object 200 built by the depth unit 120 corresponds the shooting result 20 a of the profile 202 of the object 200. As shown in FIG. 3B, the first depth image 30 b of the object 200 built by the depth unit 120 corresponds the shooting result 20 b of the profile 204 of the object 200. As shown in FIG. 3C, the first depth image 30 c of the object 200 built by the depth unit 120 corresponds the shooting result 20 c of the front side 206 of the object 200.
The image grouping unit 130 is coupled to the depth unit 120. The image grouping unit 130 divides the first depth images of the object 200 into a plurality of depth image groups according to linking information of the object 200, where the linking information records location information corresponding to a plurality of substructures of the object 200, each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups.
It should be noticed that the object 200 of the present exemplary embodiment may have a plurality of substructures, where each of the substructures is, for example, a local block of the object 200. Moreover, each of the substructures may have a plurality of microstructures, where each of the macrostructures can be regarded as a structure unit of the object 200 that cannot be further divided. Therefore, the linking information of the object 200 records a connection relationship of the substructures of the object 200 and records location information of the microstructures of each of the substructures.
According to another aspect, the object 200 of the present exemplary embodiment may have a global coordinate system, and the location information of each of the substructures on the global coordinate system is referred to as first coordinate information of the substructure, namely, the first coordinate information of each of the substructures is obtained according to the global coordinate system where the object 200 locates. Moreover, the location information of the microstructures of each of the substructures is also on such total coordinate system. In this way, since each of the substructures has the first coordinate information on the global coordinate system, the image grouping unit 130 can learn the connection relationship between the substructures according to the linking information, and a location of each of the substructures on the object 200 and location information thereof.
FIG. 4 is a schematic diagram of the linking information of the object 200 according to an exemplary embodiment of the disclosure, where a situation that the object 200 is a human body is taken as an example to describe the linking information of the present exemplary embodiment. Referring to FIG. 4, the object 200 has substructures p1˜p11, where the substructures p1-p11 have at least one or a plurality of microstructures p1-1˜p1-N11, p2-1˜p2-N2, . . . , p11-1˜p11-N11, where N1˜N11 are respectively positive integers. Moreover, each of the substructures p1˜p11 can be regarded as a rigid body, namely, during a process that the image capturing unit 110 shoots the object 200, although the object 200 can conduct the dynamic motion, the microstructures p1-1˜p1-N1, p2-1˜p2-N2, . . . , p11-1˜p11-N11 in the substructures p1˜p11 may respectively have a same moving speed, namely, each of the substructures p1˜p11 is regarded to have a single speed.
Here, the object 200 is, for example, the human body, and each of the substructures p1-p11 is, for example, a local skeleton structure of the human body, and the microstructures p1-1˜p1-N1, p2-1˜p2-N2, . . . , p11-1˜p11-N11 in the substructures p1˜p11, for example, corresponds to each bone of the human body (i.e. the aforementioned structure unit). In detail, each of the substructures p1˜p11 is, for example, a skeleton structure of a head of the human body, a skeleton structure of a front arm or upper arm, a skeleton structure of a chest, a skeleton structure of an abdomen, or a skeleton structure of a thigh or lower leg, though the disclosure is not limited thereto. On the other hand, the substructure p3 is for example, the front arm, and the microstructures p3-1˜p3-N3 of the substructure p3 are, for example, a radius bone, an ulna and a finger bone, etc. Namely, each of the substructures p1˜p11 of the object 200 can be regarded as a local skeleton structure having a plurality of bones. Therefore, in the present exemplary embodiment, the linking information records location information of a part of the microstructures (for example, bones) in the object 200, so as to define the microstructures belonging to one substructure (for example, the local skeleton structure), and according to the local information of each of the microstructures, the linking information may also include the connection relationship between the substructures and the location information of each of the substructures in the object 200 (i.e. the global coordinate system of the object 200). In this way, the object 200 can be divided into the substructures p1˜p11 according to the linking information.
It should be noticed that in the present exemplary embodiment, the structure recognizing unit 135 can be used to produce the substructures p1˜p11 of the object 200, where the structure recognizing unit 135 can be respectively coupled to the image grouping unit 130 and the depth unit 120. Further, the structure recognizing unit 135 can recognize positions of each of the microstructures p1-1˜p1-N1, . . . , p11-1˜p11-N11 of the object 200 in each of the first depth images 30 a, 30 b and 30 c according to a depth value of each pixel in the first depth images 30 a, 30 b and 30 c of the object and the location information of the microstructures p1-1˜p1-N1, . . . , p11-1˜p11-N11 of the object 200, so as to recognize an appearance of the object 200 varied along with the time. The structure recognizing unit 135 can estimate the positions of the pixels corresponding to the microstructures p1-1˜p1-N1, . . . , p11-1˜p11-N11 of the object 200 according to positions of the pixels in the first depth images 30 a, 30 b and 30 c. Therefore, the structure recognizing unit 135 can divide the substructures p1˜p11 of the object 200 by calculating the corresponding relationship between the pixels and the microstructures p1-1˜p1-N1, . . . , p11-1˜p11-N11 in the first depth images 30 a, 30 b and 30 c. In this way, the structure recognizing unit 135 can detect and track the appearance of the object 200 with a variable motion pattern by taking the microstructures p1-1˜p1-N1, . . . , p11-1˜p11-N11 as a sample and using the connection relationship between the substructures p1˜p11 in the object 200.
For example, it is also assumed that the object 200 is the human body, if the pixel in the first depth images 30 a, 30 b and 30 c is located at an edge of the first depth images 30 a, 30 b and 30 c, the structure recognizing unit 135 recognizes that the pixel is located at an extremity of the four limbs. Alternatively, if the pixel in the first depth images 30 a, 30 b and 30 c is located at a center area of the first depth images 30 a, 30 b and 30 c, the structure recognizing unit 135 recognizes that the pixel is located at a torso portion of the object 200. Moreover, if the pixel in the first depth images 30 a, 30 b and 30 c is located close to a region where the microstructures of the object 200 locate, the structure recognizing unit 135 recognizes that the pixel is located at a bone area of the object 200. In this way, according to the first depth images 30 a, 30 b and 30 c of the object 200 and the location information of the microstructures p1-1˜p1-N1, . . . , p11-1˜p11-N11 of the object 200, the structure recognizing unit 135 can use the first depth images 30 a, 30 b and 30 to recognize the microstructures p1-1˜p1-N1, . . . , p11-1˜p11-N11 of the object 200, and track the appearance of the object 200. Therefore, the structure recognizing unit 135 can recognize the microstructures that have the same speed during the process that the image capturing unit 110 shoots the object 200, so as to find the substructures composed of the microstructures.
Moreover, the image grouping unit 130 divides each of the first depth images 30 a, 30 b and 30 c of the object 200 into a plurality of second depth images according to the substructures p1˜p11 of the object 200 and the depth values of the first depth images 30 a, 30 b and 30 c, and takes the corresponding second depth images in the first depth images 30 a, 30 b and 30 c as an image group, so as to divide depth image groups located at local blocks of each of the substructures p1˜p11 of the object 200.
FIGS. 5A-5B are schematic diagrams of a situation that the image grouping unit 130 divides the first depth images of the object 200 into a plurality of depth image groups according to an exemplary embodiment of the disclosure, which are described below with reference of the first depth images 30 a, 30 b and 30 c of FIGS. 3A-3B and the substructures of FIG. 4.
Referring to FIG. 5A and FIG. 5B, according to the linking information of the object 200, the image grouping unit 130 divides each of the first depth images 30 a, 30 b and 30 c of the object 200 into a plurality of second depth images 501˜511 according to the divided substructures p1˜p11 of the object 200 and the depth values of the first depth images 30 a, 30 b and 30 c, and takes the corresponding second depth images 501˜511 in the first depth images 30 a, 30 b and 30 c as an image group, so as to divide a plurality of depth image groups 501 s˜511 s.
In detail, as shown in FIG. 5A, regarding the first depth image 30 a, the image grouping unit 130 divides the first depth image 30 a into the second depth images 501˜511 according to the substructures p1˜p11 of the linking information of the object 200, for example, depth images corresponding to the head, the front arm or upper arm, the chest, the abdomen, the thigh or lower leg of the human body. Similarly, the image grouping unit 130 divides first depth images 30 b and 30 c into the second depth images 501˜511 the according to the substructures p1˜p11 of the linking information of the object 200. Then, as shown in FIG. 5B, taking the second depth image 501 (i.e. the depth image of the head of the object 200) as an example, the image grouping unit 130 takes the second depth images 501 of the first depth images 30 a, 30 b and 30 c as a depth image group, and the other second depth images 502˜511 can be deduced by analogy. In this way, the image grouping unit 130 generates the depth image groups 501 s˜511 s on the substructures p1˜p11.
The model building unit 140 is coupled to the image grouping unit 130. The model building unit 140 builds a local model corresponding to each of the substructures p1˜p11 on the object 200 according to the second depth images 501˜511 and the location information corresponding to each of the substructures p1˜p11. Moreover, the model building unit 140 merges the local models corresponding to the substructures p1˜p11 according to the linking information of the object 200, so as to build the 3D model of the object 200.
FIG. 5C is a schematic diagram illustrating a situation that the model building unit 140 builds a local model corresponding to each of the substructures according to an exemplary embodiment of the disclosure. FIG. 5D is a schematic diagram illustrating a situation that the model building unit 140 merges the local models to build the 3D model of the object 200 according to an exemplary embodiment of the disclosure.
Referring to FIGS. 5B-5D, according to local coordinate systems C1˜C11 of the substructures p1˜p11 (shown in FIG. 4), the model building unit 140 obtains location information and orientation information of the microstructures in each of the substructures p1˜p11 relative to the local coordinate systems C1˜C11, and respectively generates local models M1˜M11 of the object 200 according to the depth image groups 501 s˜511 s corresponding to the substructures p1˜p11. For example, it is also assumed that the object 200 is the human body, the local model M1 corresponds to the head, the local models M2˜M3, M5˜M6 correspond to the front arm and the upper arms, the local model M4 corresponds to the chest, the local model M7 corresponds to the abdomen, and the local models M8˜M11 correspond to thighs and the lower legs.
Further, the 3D model reconstructing apparatus 10 of the present exemplary embodiment further selectively includes the coordinate transforming unit 150. The coordinate transforming unit 150 is coupled to the image grouping unit 130 and the model building unit 140. Here, the coordinate transforming unit 150 transforms the first coordinate information of each of the substructures from the global coordinate system Co of the object 200 to the local coordinate systems C1˜C11 (for example, cylindrical coordinate systems) of the substructures p1˜p11 through a transforming calculation, where the local coordinate systems C1˜C11 correspond to the substructures p1˜p11, and the substructures p1˜p11 have the second coordinate information in the local coordinate systems C1˜C11. Namely, the substructures p1˜p11 of the object 200 respectively have their own local coordinate systems C1˜C11, and each of the substructures p1˜p11 have location information of the local coordinate systems C1˜C11, which is referred to as second coordinate information.
According to another aspect, since the first depth images built by the depth unit 120 according to the shooting results provided by the image capturing unit 110 have the first coordinate information of the global coordinate system Co, and during the process that the image capturing unit 110 shoots the object 200, each of the substructures and the microstructures in each of the substructures can be regarded as a rigid body, namely, the appearance of each of the microstructures is not changed (for example, the appearance of the of the microstructures is not elongated, shortened or distorted) during the shooting process of the image capturing unit 110. Therefore, regarding the first depth image of the object 200 generated corresponding to each of the shooting angles, the 3D coordinate information of the pixels in the depth image can be stored according to the respective local coordinate system of each of the substructures of the object 200. In this way, the model building unit 140 can merge the depth image groups 501 s˜511 s corresponding to the substructures p1˜p11 according to the second coordinate information of the local coordinate systems C1˜C11, so as to build the local models M1˜M11 corresponding to each of the substructures p1˜p11.
Moreover, when the model building unit 140 builds the local models M1˜M11 corresponding to each of the substructures p1˜p11, the model building unit 140 further determines whether each of the second depth images 501˜511 in the depth image groups 501 s˜511 s have an overlapped region. If the second depth images 501˜511 have the overlapped region, the model building unit 140 excludes the overlapped region, and builds the local models according to remained depth images in the second depth images 501-511 excluding the overlapped region. For example, regarding the depth image group 501 s of FIG. 5C, if the three second depth images 501 of the depth image group 501 s have an overlapped region, the model building unit 140 excludes the overlapped region, and uses the remained depth images of the second depth images 501 excluding the overlapped region to build the local models M1˜M11 of the object 200 in the substructure p1.
In detail, when the model building unit 140 builds the local models M1˜M11 corresponding to the substructures p1˜p11, each of the second depth images 501˜511 in the depth image groups 501 s˜511 s have an overlapped region, when the model building unit 140 merges the second depth images 501˜511 in each of the depth image groups, the model building unit 140 adjusts a configuration of each of the second depth images 501˜511 of the depth image groups 501 s˜511 s on each of the local coordinate systems according to feature information of the object 200 on the surface of each of the substructures p1˜p11, so as to build the local models M1˜M11 corresponding to the substructures p1˜p11. The feature information is, for example, a surface texture of the object 200, or a profile of the object 200, etc., which is not limited by the disclosure.
In other words, in order to ensure that each of the substructures p1˜p11 accurately corresponds to each of the second depth images in each of the depth image groups, for example, the skeleton structure of the head (including a position and a direction of each skull) accurately corresponds to the depth image of the head, or the skeleton structure of the arm (including a position and direction of each hand bone) accurately corresponds to the depth image of the arm, etc., the model building unit 140 further finely adjusts the partial models M1˜M11 according to the features of the surface texture or the profile, etc. of the object 200. In this way, the model building unit 140 can minimize the overlapped region between the second depth images in each of the depth image groups, so as to build the accurate local models M1˜M11.
For example, FIGS. 5E-5G are schematic diagrams of local models of the object 200 built by the model building unit 140 viewed from different viewing-angles according to an exemplary embodiment of the disclosure. The local model M4 of the object 200 in FIG. 5D is taken as an example for descriptions, and the local model M4 is a chest model of the object 200. Referring to FIGS. 5E-5G, the local model M4 is generated by the model building unit 140 according to the depth image group 504 s. As shown in FIGS. 5E-5G, the local model M4 can be composed of a second depth image 504 a (coming from the first depth image 30 a), a second depth image 504 b (coming from the first depth image 30 b), and a second depth image 504 c (coming from the first depth image 30 c).
Referring to FIG. 5D, after the model building unit 140 builds the local models M1˜M11 corresponding to the substructures, the model building unit 140 further merges the local models M1˜M11 on the local coordinate systems of the substructures according to the linking information of the object 200 to transform the local models M1˜M11 of the substructures from the local coordinate systems to the global coordinate system Co of the object 200, so as to merge the local models M1˜M11 corresponding to the substructures to build the 3D model of the object 200. Namely, the model building unit 140 transforms the depth image groups corresponding to the substructures from the local coordinate systems to the global coordinate system Co of the object 200 according to the second coordinate information of the substructures on the local coordinate systems, and merges the local models M1˜M11 to generate a 3D model Mf of the integral object 200.
It should be noticed that, the model building unit 140 further merges the local models M1˜M11 according to the connection relationship between the microstructures of each of the substructures. In detail, the model building unit 140 takes each pixel in the second depth image corresponding to each of the substructures as a point, and during a process that the model building unit 140 merges the local models M1˜M11, and when three neighboring points are connected to another point, the model building unit 140 determines whether a first figure formed by the three neighboring points is overlapped with a second figure formed by two of the three neighboring points and the other point. In detail, the first figure and the second figure are, for example, respectively a triangle, the model building unit 140 takes a triangle formed by each point and two neighboring points as a unit for connection (i.e. form a plurality of triangles), and determines whether an overlapping phenomenon occurs between the connected triangles, i.e. checks whether the 3D planes represented by the triangles have spatial relationships of plane intersection and plane overlapping. If the 3D planes represented by the triangles have spatial relationships of plane intersection and plane overlapping, such connection relationship is incorrect, and the model building unit 140 selects another point to execute the above determination and check operations. If the overlapping phenomenon is not occurred, the model building unit 140 determines that the connection relationship between the checked three points is correct.
FIG. 5H is a schematic diagram of the connection relationship of the 3D model Mf built by the model building unit 140 according to an exemplary embodiment of the disclosure. Referring to FIG. 5H, the model building unit 140 determines whether the triangles formed by each three points in the points Q1-Q4 have the overlapping phenomenon. For example, regarding a triangle T1 formed by the point Q1 and the neighboring points Q3 and Q4, and a triangle T2 formed by the point Q2 and the neighboring points Q3 and Q4, since the overlapping phenomenon occurs between the triangle T1 and the triangle T2, the model building unit 140 determines that the connection relationship of the triangle T1 and the triangle T2 (i.e. the connection relationship of the points Q1, Q3 and Q4 and the connection relationship of the points Q2, Q3 and Q4) is incorrect.
On the other hand, regarding the triangle T1 formed by the point Q1 and the neighboring points Q3 and Q4, and a triangle T2′ formed by the point Q2 and the neighboring points Q1 and Q4, the connection relationship of the triangle T1 and the triangle T2′ (i.e. the connection relationship of the points Q1, Q3 and Q4 and the connection relationship of the points Q2, Q1 and Q4) is correct. In this way, the model building unit 140 checks whether a triangle formed by each three points is overlapped with the neighboring triangles until all of the points and the neighboring points have the connection relationships.
Moreover, in the present exemplary embodiment, the display unit 160 can be selectively configured for coupling to the model building unit 140. The display unit 160 can display the 3D model of the object 200 built by the model building unit 140. In this way, a user can view the accurate 3D model of the object 200 through the display unit 160.
According to the above descriptions, the 3D model reconstructing apparatus 10 of the present exemplary embodiment obtains shooting information of the object 200 in different angles through the image capturing unit 110, and generates a plurality of the first depth images 30 a-30 c of the object 200 through the depth unit 120. Moreover, the 3D model reconstructing apparatus 10 divides the first depth images 30 a-30 c into a plurality of depth image groups 501 s˜511 s according to the linking information of the object 200. Particularly, the 3D model reconstructing apparatus 10 generates the local models M1˜M11 according to the coordinate systems of the depth image groups 501 s˜511 s. Finally, the 3D model reconstructing apparatus 10 merges the local models M1˜M11 to generate the 3D model Mf of the object 200. In this way, according to the present exemplary embodiment, an accurate 3D model of a deformable object can be built through the image capturing unit 110.
FIG. 6 is a flowchart illustrating a method for reconstructing a 3D model according to an exemplary embodiment of the disclosure.
Referring to FIG. 6, in step S602, the image capturing unit 110 is used to capture shooting information of the object 200 in different angles to provide a plurality of shooting results. In step S604, the depth unit 120 builds a plurality of first depth images of the object 200 according to the shooting results provided by the image capturing unit 110, where the first depth image has depth information of the object 200.
In step S606, the structure recognizing unit 135 calculates a corresponding relationship between a plurality of microstructures in the object 200 and the first depth images according to linking information of the object 200, so as to divide a plurality of substructures of the object 200. Then, in step S608, the image grouping unit 130 divides the first depth images of the object 200 into a plurality of depth image groups according to the linking information and the substructures of the object 200, where the linking information records location information corresponding to a plurality of substructures of the object 200, each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups.
In step S610, the model building unit 140 determines whether an overlapping region exists between the second depth images of each of the depth image groups.
If the model building unit 140 determines that none overlapping region exists between the second depth images of each of the depth image groups, a step S612 is executed, by which the model building unit 140 builds a local model corresponding to each of the substructures according to the second depth images and the location information corresponding to each of the substructures.
However, if the model building unit 140 determines that the overlapping region exists between the second depth images of each of the depth image groups, a step S614 is executed, by which the model building unit 140 excludes the overlapping region, and builds the local model according to remained depth images in the second depth images excluding the overlapped region by using feature information of the object 200 on the surface of each of the substructures.
In step S616, the model building unit 140 transforms the local models of the substructures from the local coordinate systems to the global coordinate system of the object 200. Finally, in step S618, the model building unit 140 merges the local models on the local coordinate systems of the substructures according to the linking information of the object 200, so as to build the 3D model of the object 200.
In summary, according to the method for reconstructing a 3D model and the 3D model reconstructing apparatus of the disclosure, the first depth images of the object in different angles are divided into a plurality of depth image groups according to the linking information of the object, where each of the first depth image groups has a plurality of second depth images. Moreover, each of the depth image groups is transformed from the global coordinate system of the object to the respective local coordinate systems of the depth image groups, so as to generate the local model by using each of the depth image groups. Then, an integral 3D model of the object 200 is built by merging the local models. In this way, the 3D image of the deformable object is accurately built.
It will be apparent to those skilled in the art that various modifications and variations can be made to the disclosed embodiments. It is intended that the specification and examples be considered as exemplary only, with a true scope of the disclosure being indicated by the following claims and their equivalents.

Claims (18)

What is claimed is:
1. A method for reconstructing a three-dimensional model, adapted to build a three-dimensional model of an object, comprising:
obtaining a plurality of first depth images of the object, wherein the object is a human body;
dividing the first depth images of the object into a plurality of depth image groups according to linking information of the object, wherein the linking information records location information corresponding to a plurality of substructures of the object, each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups;
building a local model corresponding to each of the substructures according to the second depth images and the location information corresponding to each of the substructures; and
merging the local models corresponding to the substructures according to the linking information of the object, so as to build the three-dimensional model of the object,
adjusting the local models according to the features of the surface of the object,
wherein the step of merging the local models corresponding to the substructures according to the linking information of the object to build the three-dimensional model of the object further comprises:
in a plurality of pixels in each of the second depth images corresponding to each of the substructures, when three neighboring pixels are connected to another pixel, determining whether a first figure formed by the three neighboring pixels is overlapped with a second figure formed by two of the three neighboring pixels and the other pixel;
when the first figure is overlapped with the second figure, determining a connection relationship between the three neighboring pixels and the other pixel to be incorrect; and
when the first figure is not overlapped with the second figure, maintaining the connection between the three neighboring pixels and the other pixel.
2. The method for reconstructing the three-dimensional model as claimed in claim 1, wherein the step of dividing the first depth images of the object into the depth image groups according to the linking information of the object comprises:
dividing the substructures of the object according to a corresponding relationship between a plurality of pixels in the first depth images and a plurality of microstructures of the object.
3. The method for reconstructing the three-dimensional model as claimed in claim 1, wherein the location information respectively corresponding to the substructures has first coordinate information of each of the substructures, and the first coordinate information is obtained according to a global coordinate system of the object.
4. The method for reconstructing the three-dimensional model as claimed in claim 3, wherein the step of building the local model corresponding to each of the substructures according to the second depth images and the location information corresponding to each of the substructures comprises:
transforming the first coordinate information corresponding to each of the substructures and the second depth images corresponding to the substructures from the global coordinate system to a local coordinate system through a transforming calculation, wherein the local coordinate system corresponds to the substructure, and the substructure has second coordinate information in the local coordinate system.
5. The method for reconstructing the three-dimensional model as claimed in claim 4, wherein after the step of transforming the first coordinate information corresponding to each of the substructures and the second depth images corresponding to each of the substructures from the global coordinate system to the local coordinate system, the method further comprises:
obtaining location information and orientation information of at least one microstructure in each of the substructures relative to the local coordinate system according to the second coordinate information of the substructure.
6. The method for reconstructing the three-dimensional model as claimed in claim 4, wherein the step of merging the local models corresponding to the substructures according to the linking information of the object to build the three-dimensional model of the object comprises:
transforming the second coordinate information in the local model corresponding to each of the substructures from the local coordinate system to the global coordinate system of the object; and
merging the local models of the substructures on the global coordinate system according to the linking information of the object, so as to build the three-dimensional model of the object.
7. The method for reconstructing the three-dimensional model as claimed in claim 1, wherein the step of building the local model corresponding to each of the substructures according to the second depth images and the location information corresponding to each of the substructures comprises:
obtaining an overlapping region between the second depth images in each of the depth image groups, and excluding the overlapping region to build the local model.
8. The method for reconstructing the three-dimensional model as claimed in claim 1, wherein the step of building the local model corresponding to each of the substructures according to the second depth images and the location information corresponding to each of the substructures further comprises:
building the local model according to feature information of the substructure.
9. The method for reconstructing the three-dimensional model as claimed in claim 1, further comprising displaying the three-dimensional model of the object.
10. A three-dimensional model reconstructing apparatus, adapted to build a three-dimensional model of an object, the three-dimensional model reconstructing apparatus comprising:
a camera, capturing shooting information of the object in different angles to provide a plurality of shooting results, wherein the object is a human body;
a processor electrically connected to the camera and configured to execute:
a depth unit to build a plurality of first depth images of the object according to the shooting results provided by the camera;
an image grouping unit to divide the first depth images of the object into a plurality of depth image groups according to linking information of the object, wherein the linking information records location information corresponding to a plurality of substructures of the object, each of the depth image groups comprises a plurality of second depth images, and each of the substructures corresponds to the second depth images of each of the depth image groups; and
a model building unit to build a local model corresponding to each of the substructures according to the second depth images and the location information corresponding to each of the substructures, and merge the local models corresponding to the substructures according to the linking information of the object, so as to build the three-dimensional model of the object, and further adjust the local models according to the features of the surface of the object,
wherein in a plurality of pixels in each of the second depth images corresponding to each of the substructures, when three neighboring pixels are connected to another pixel, the model building unit determines whether a first figure formed by the three neighboring pixels is overlapped with a second figure formed by two of the three neighboring pixels and the other pixel, when the first figure is overlapped with the second figure, the model building unit determines a connection relationship between the three neighboring pixels and the other pixel to be incorrect, and when the first figure is not overlapped with the second figure, the model building unit maintains the connection between the three neighboring pixels and the other pixel.
11. The three-dimensional model reconstructing apparatus as claimed in claim 10, wherein the processor is further configured to execute:
a structure recognizing unit to divide the substructures of the object according to a corresponding relationship between a plurality of pixels in the first depth images and a plurality of microstructures of the object.
12. The three-dimensional model reconstructing apparatus as claimed in claim 10, wherein the location information respectively corresponding to the substructures has first coordinate information of each of the substructures, and the first coordinate information is obtained according to a global coordinate system of the object.
13. The three-dimensional model reconstructing apparatus as claimed in claim 12, wherein the processor is further configured to execute:
a coordinate transforming unit to transform the first coordinate information corresponding to each of the substructures and the second depth images corresponding to the substructures from the global coordinate system to a local coordinate system through a transforming calculation, wherein the local coordinate system corresponds to the substructure, and the substructure has second coordinate information in the local coordinate system.
14. The three-dimensional model reconstructing apparatus as claimed in claim 13, wherein the model building unit obtains location information and orientation information of at least one microstructure in each of the substructures relative to the local coordinate system according to the second coordinate information of the substructure.
15. The three-dimensional model reconstructing apparatus as claimed in claim 13, wherein the model building unit transforms the second coordinate information in the local model corresponding to each of the substructures from the local coordinate system to the global coordinate system of the object, and merges the local models of the substructures on the global coordinate system according to the linking information of the object, so as to build the three-dimensional model of the object.
16. The three-dimensional model reconstructing apparatus as claimed in claim 10, wherein the model building unit obtains an overlapping region between the second depth images in each of the depth image groups, and excludes the overlapping region to build the local model.
17. The three-dimensional model reconstructing apparatus as claimed in claim 10, wherein the model building unit builds the local model according to feature information of the substructure.
18. The three-dimensional model reconstructing apparatus as claimed in claim 10, further comprising:
a display, coupled to the processor, and displaying the three-dimensional model of the object.
US13/686,927 2012-10-04 2012-11-28 Method and apparatus for reconstructing three dimensional model Active 2034-02-01 US9262862B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
TW101136721A 2012-10-04
TW101136721 2012-10-04
TW101136721A TWI466062B (en) 2012-10-04 2012-10-04 Method and apparatus for reconstructing three dimensional model

Publications (2)

Publication Number Publication Date
US20140099017A1 US20140099017A1 (en) 2014-04-10
US9262862B2 true US9262862B2 (en) 2016-02-16

Family

ID=50432704

Family Applications (1)

Application Number Title Priority Date Filing Date
US13/686,927 Active 2034-02-01 US9262862B2 (en) 2012-10-04 2012-11-28 Method and apparatus for reconstructing three dimensional model

Country Status (2)

Country Link
US (1) US9262862B2 (en)
TW (1) TWI466062B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10922894B2 (en) * 2016-06-06 2021-02-16 Biodigital, Inc. Methodology and system for mapping a virtual human body
US11741670B2 (en) 2021-03-01 2023-08-29 Samsung Electronics Co., Ltd. Object mesh based on a depth image

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9829984B2 (en) * 2013-05-23 2017-11-28 Fastvdo Llc Motion-assisted visual language for human computer interfaces
US10146299B2 (en) * 2013-11-08 2018-12-04 Qualcomm Technologies, Inc. Face tracking for additional modalities in spatial interaction
US20150243035A1 (en) * 2014-02-21 2015-08-27 Metaio Gmbh Method and device for determining a transformation between an image coordinate system and an object coordinate system associated with an object of interest
TW201719572A (en) * 2015-11-19 2017-06-01 國立交通大學 Method for analyzing and searching 3D models
JP6333871B2 (en) * 2016-02-25 2018-05-30 ファナック株式会社 Image processing apparatus for displaying an object detected from an input image
TWI567691B (en) * 2016-03-07 2017-01-21 粉迷科技股份有限公司 Method and system for editing scene in three-dimensional space
CN109559371B (en) * 2017-09-27 2023-12-26 虹软科技股份有限公司 Method and device for three-dimensional reconstruction
CN108053437B (en) * 2017-11-29 2021-08-03 奥比中光科技集团股份有限公司 Three-dimensional model obtaining method and device based on posture
US10643342B2 (en) * 2018-02-08 2020-05-05 Huawei Technologies Co., Ltd. Group optimization depth information method and system for constructing a 3D feature map
JP7119425B2 (en) * 2018-03-01 2022-08-17 ソニーグループ株式会社 Image processing device, encoding device, decoding device, image processing method, program, encoding method and decoding method
US10762219B2 (en) * 2018-05-18 2020-09-01 Microsoft Technology Licensing, Llc Automatic permissions for virtual objects
US10846923B2 (en) * 2018-05-24 2020-11-24 Microsoft Technology Licensing, Llc Fusion of depth images into global volumes
CN108961388B (en) * 2018-06-05 2023-03-21 哈尔滨工业大学深圳研究生院 Microstructure three-dimensional modeling method, microstructure three-dimensional modeling device, microstructure three-dimensional modeling equipment and computer storage medium
DE102018113580A1 (en) * 2018-06-07 2019-12-12 Christoph Karl METHOD AND DEVICE FOR PRODUCING AN IMPLANT
CN109712230B (en) * 2018-11-27 2023-02-28 先临三维科技股份有限公司 Three-dimensional model supplementing method and device, storage medium and processor
US11276250B2 (en) * 2019-10-23 2022-03-15 International Business Machines Corporation Recognition for overlapped patterns
AU2022328457A1 (en) * 2021-08-18 2024-03-07 mPort Ltd Methods for generating a partial three-dimensional representation of a person

Citations (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6249600B1 (en) 1997-11-07 2001-06-19 The Trustees Of Columbia University In The City Of New York System and method for generation of a three-dimensional solid model
US20030038801A1 (en) * 2001-08-24 2003-02-27 Sanyo Electric Co., Ltd. Three dimensional modeling apparatus
US20030063085A1 (en) * 2001-10-02 2003-04-03 Leow Wee Kheng Frontier advancing polygonization
US6754370B1 (en) 2000-08-14 2004-06-22 The Board Of Trustees Of The Leland Stanford Junior University Real-time structured light range scanning of moving scenes
US7003134B1 (en) 1999-03-08 2006-02-21 Vulcan Patents Llc Three dimensional object pose estimation which employs dense depth information
US7010158B2 (en) 2001-11-13 2006-03-07 Eastman Kodak Company Method and apparatus for three-dimensional scene modeling and reconstruction
US7184071B2 (en) 2002-08-23 2007-02-27 University Of Maryland Method of three-dimensional object reconstruction from a video sequence using a generic model
US7450736B2 (en) 2005-10-28 2008-11-11 Honda Motor Co., Ltd. Monocular tracking of 3D human motion with a coordinated mixture of factor analyzers
US20090175540A1 (en) 2007-12-21 2009-07-09 Honda Motor Co., Ltd. Controlled human pose estimation from depth image streams
TWM364920U (en) 2009-04-10 2009-09-11 Shen-Jwu Su 3D human face identification device with infrared light source
US7590262B2 (en) 2003-05-29 2009-09-15 Honda Motor Co., Ltd. Visual tracking using depth data
US20100034457A1 (en) * 2006-05-11 2010-02-11 Tamir Berliner Modeling of humanoid forms from depth maps
US20100194872A1 (en) * 2009-01-30 2010-08-05 Microsoft Corporation Body scan
US20100197390A1 (en) * 2009-01-30 2010-08-05 Microsoft Corporation Pose tracking pipeline
US20100259546A1 (en) 2007-09-06 2010-10-14 Yeda Research And Development Co. Ltd. Modelization of objects in images
CN101894383A (en) 2010-06-11 2010-11-24 四川大学 Method for accelerating ray-traced digital image rebuilding technology
US7860301B2 (en) 2005-02-11 2010-12-28 Macdonald Dettwiler And Associates Inc. 3D imaging system
US7961910B2 (en) 2009-10-07 2011-06-14 Microsoft Corporation Systems and methods for tracking a model
US20110181591A1 (en) * 2006-11-20 2011-07-28 Ana Belen Benitez System and method for compositing 3d images
US7999811B2 (en) 2007-01-16 2011-08-16 Sony Corporation Image processing device, method, and program, and objective function
US20110210915A1 (en) 2009-05-01 2011-09-01 Microsoft Corporation Human Body Pose Estimation
US20110293137A1 (en) * 2010-05-31 2011-12-01 Primesense Ltd. Analysis of three-dimensional scenes
TW201223248A (en) 2010-07-28 2012-06-01 Sisvel Technology Srl Method for combining images relating to a three-dimensional content
US20120147004A1 (en) 2010-12-13 2012-06-14 Electronics And Telecommunications Research Institute Apparatus and method for generating digital actor based on multiple images
US8208717B2 (en) 2009-02-25 2012-06-26 Seiko Epson Corporation Combining subcomponent models for object image modeling
US20130100119A1 (en) * 2011-10-25 2013-04-25 Microsoft Corporation Object refinement using many data sets
US20130187919A1 (en) * 2012-01-24 2013-07-25 University Of Southern California 3D Body Modeling, from a Single or Multiple 3D Cameras, in the Presence of Motion
US20130286012A1 (en) * 2012-04-25 2013-10-31 University Of Southern California 3d body modeling from one or more depth cameras in the presence of articulated motion
US8610723B2 (en) * 2011-06-22 2013-12-17 Microsoft Corporation Fully automatic dynamic articulated model calibration
US8639020B1 (en) * 2010-06-16 2014-01-28 Intel Corporation Method and system for modeling subjects from a depth map
US20140225988A1 (en) * 2011-09-07 2014-08-14 Commonwealth Scientific And Industrial Research Organisation System and method for three-dimensional surface imaging

Patent Citations (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6249600B1 (en) 1997-11-07 2001-06-19 The Trustees Of Columbia University In The City Of New York System and method for generation of a three-dimensional solid model
US7003134B1 (en) 1999-03-08 2006-02-21 Vulcan Patents Llc Three dimensional object pose estimation which employs dense depth information
US7158656B2 (en) 1999-03-08 2007-01-02 Vulcan Patents Llc Three dimensional object pose estimation which employs dense depth information
US6754370B1 (en) 2000-08-14 2004-06-22 The Board Of Trustees Of The Leland Stanford Junior University Real-time structured light range scanning of moving scenes
US20030038801A1 (en) * 2001-08-24 2003-02-27 Sanyo Electric Co., Ltd. Three dimensional modeling apparatus
US20030063085A1 (en) * 2001-10-02 2003-04-03 Leow Wee Kheng Frontier advancing polygonization
US7010158B2 (en) 2001-11-13 2006-03-07 Eastman Kodak Company Method and apparatus for three-dimensional scene modeling and reconstruction
US7184071B2 (en) 2002-08-23 2007-02-27 University Of Maryland Method of three-dimensional object reconstruction from a video sequence using a generic model
US7590262B2 (en) 2003-05-29 2009-09-15 Honda Motor Co., Ltd. Visual tracking using depth data
US7860301B2 (en) 2005-02-11 2010-12-28 Macdonald Dettwiler And Associates Inc. 3D imaging system
US7450736B2 (en) 2005-10-28 2008-11-11 Honda Motor Co., Ltd. Monocular tracking of 3D human motion with a coordinated mixture of factor analyzers
US20100034457A1 (en) * 2006-05-11 2010-02-11 Tamir Berliner Modeling of humanoid forms from depth maps
US20110181591A1 (en) * 2006-11-20 2011-07-28 Ana Belen Benitez System and method for compositing 3d images
US7999811B2 (en) 2007-01-16 2011-08-16 Sony Corporation Image processing device, method, and program, and objective function
US20100259546A1 (en) 2007-09-06 2010-10-14 Yeda Research And Development Co. Ltd. Modelization of objects in images
US20090175540A1 (en) 2007-12-21 2009-07-09 Honda Motor Co., Ltd. Controlled human pose estimation from depth image streams
US20100194872A1 (en) * 2009-01-30 2010-08-05 Microsoft Corporation Body scan
US20100197390A1 (en) * 2009-01-30 2010-08-05 Microsoft Corporation Pose tracking pipeline
US8009867B2 (en) 2009-01-30 2011-08-30 Microsoft Corporation Body scan
US8208717B2 (en) 2009-02-25 2012-06-26 Seiko Epson Corporation Combining subcomponent models for object image modeling
TWM364920U (en) 2009-04-10 2009-09-11 Shen-Jwu Su 3D human face identification device with infrared light source
US20110210915A1 (en) 2009-05-01 2011-09-01 Microsoft Corporation Human Body Pose Estimation
US7961910B2 (en) 2009-10-07 2011-06-14 Microsoft Corporation Systems and methods for tracking a model
US20110293137A1 (en) * 2010-05-31 2011-12-01 Primesense Ltd. Analysis of three-dimensional scenes
CN101894383A (en) 2010-06-11 2010-11-24 四川大学 Method for accelerating ray-traced digital image rebuilding technology
US8639020B1 (en) * 2010-06-16 2014-01-28 Intel Corporation Method and system for modeling subjects from a depth map
TW201223248A (en) 2010-07-28 2012-06-01 Sisvel Technology Srl Method for combining images relating to a three-dimensional content
US20120147004A1 (en) 2010-12-13 2012-06-14 Electronics And Telecommunications Research Institute Apparatus and method for generating digital actor based on multiple images
US8610723B2 (en) * 2011-06-22 2013-12-17 Microsoft Corporation Fully automatic dynamic articulated model calibration
US20140225988A1 (en) * 2011-09-07 2014-08-14 Commonwealth Scientific And Industrial Research Organisation System and method for three-dimensional surface imaging
US20130100119A1 (en) * 2011-10-25 2013-04-25 Microsoft Corporation Object refinement using many data sets
US20130187919A1 (en) * 2012-01-24 2013-07-25 University Of Southern California 3D Body Modeling, from a Single or Multiple 3D Cameras, in the Presence of Motion
US20130286012A1 (en) * 2012-04-25 2013-10-31 University Of Southern California 3d body modeling from one or more depth cameras in the presence of articulated motion

Non-Patent Citations (12)

* Cited by examiner, † Cited by third party
Title
"Notice of Allowance of Taiwan Counterpart Application", issued on Oct. 29, 2014, p. 1-p. 4.
Allen et al., "Articulated Body Deformation from Range Scan Data," Proceedings of ACM SIGGRAPH 2002 21(3), Jul. 2002, pp. 612-619.
Baran et al., "Automatic Rigging and Animation of 3D Characters," Proceedings of ACM SIGGRAPH 2007 26(3), Jul. 2007, pp. 1-8.
Chang et al., "Global Registration of Dynamic Range Scans for Articulated Model Reconstruction," ACM Transactions on Graphics 30(3), May 2011, Article 26, pp. 26:1-26:15.
Chang, Will, and Matthias Zwicker. "Global registration of dynamic range scans for articulated model reconstruction." ACM Transactions on Graphics (TOG) 30.3 (2011): 26. *
Fitzpatrick, Richard. "Coordinate Transformations." Newtonian Dynamics. The University of Texas at Austin, Mar. 31, 2011. Web. Feb. 5, 2015. *
Pekelny et al., "Articulated Object Reconstruction and Markerless Motion Capture from Depth Video," Eurographics 2008 27(2), Apr. 2008, pp. 399-408.
Rutishauser, Martin, Markus Stricker, and Marjan Trobina. "Merging range images of arbitrarily shaped objects." Computer Vision and Pattern Recognition, 1994. Proceedings CVPR'94., 1994 IEEE Computer Society Conference on. IEEE, 1994. *
Shotton et al., "Real-Time Human Pose Recognition in Parts from a Single Depth Image", 2011 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Jun. 20-25, 2011, pp. 1297-1304.
Tong et al., "Scanning 3D Full Human Bodies Using Kinects," IEEE Transactions on Visualization and Computer Graphics 18(4), Apr. 2012, pp. 643-650.
Tong, Jing, et al. "Scanning 3d full human bodies using kinects." Visualization and Computer Graphics, IEEE Transactions on 18.4 (2012): 643-650. *
Weisstein, Eric W. "Point-Plane Distance." From MathWorld-A Wolfram Web Resource. http://mathworld.wolfram.com/Point-PlaneDistance.html. *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10922894B2 (en) * 2016-06-06 2021-02-16 Biodigital, Inc. Methodology and system for mapping a virtual human body
US11741670B2 (en) 2021-03-01 2023-08-29 Samsung Electronics Co., Ltd. Object mesh based on a depth image

Also Published As

Publication number Publication date
TW201415413A (en) 2014-04-16
TWI466062B (en) 2014-12-21
US20140099017A1 (en) 2014-04-10

Similar Documents

Publication Publication Date Title
US9262862B2 (en) Method and apparatus for reconstructing three dimensional model
CN111060023B (en) High-precision 3D information acquisition equipment and method
US9898651B2 (en) Upper-body skeleton extraction from depth maps
Sugano et al. Learning-by-synthesis for appearance-based 3d gaze estimation
US11263443B2 (en) Centimeter human skeleton pose estimation
CN109840500B (en) Three-dimensional human body posture information detection method and device
US20150347833A1 (en) Noncontact Biometrics with Small Footprint
CN109903376B (en) Face geometric information assisted three-dimensional face modeling method and system
CN102971768B (en) Posture state estimation unit and posture state method of estimation
US11398044B2 (en) Method for face modeling and related products
US20140111507A1 (en) 3-dimensional shape reconstruction device using depth image and color image and the method
WO2014154839A1 (en) High-definition 3d camera device
CN103839277A (en) Mobile augmented reality registration method of outdoor wide-range natural scene
US20120194513A1 (en) Image processing apparatus and method with three-dimensional model creation capability, and recording medium
JP2019096113A (en) Processing device, method and program relating to keypoint data
US11928778B2 (en) Method for human body model reconstruction and reconstruction system
Lim et al. Camera-based hand tracking using a mirror-based multi-view setup
JP2012181646A (en) Data processor, data processing system, and program
TW201813366A (en) Method and system for scanning an environment
Amar et al. Synthesizing reality for realistic physical behavior of virtual objects in augmented reality applications for smart-phones
Yu et al. Humbi 1.0: Human multiview behavioral imaging dataset
KR20160049639A (en) Stereoscopic image registration method based on a partial linear method
Jiménez et al. Face tracking and pose estimation with automatic three-dimensional model construction
Ortiz-Coder et al. Accurate 3d reconstruction using a videogrammetric device for heritage scenarios
TWI524213B (en) Controlling method and electronic apparatus

Legal Events

Date Code Title Description
AS Assignment

Owner name: INDUSTRIAL TECHNOLOGY RESEARCH INSTITUTE, TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TSAI, YAO-YANG;TENN, HIAN-KUN;HUANG, JAY;REEL/FRAME:029378/0062

Effective date: 20121115

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8