WO2009089129A1 - Global camera path optimization - Google Patents

Global camera path optimization Download PDF

Info

Publication number
WO2009089129A1
WO2009089129A1 PCT/US2009/030068 US2009030068W WO2009089129A1 WO 2009089129 A1 WO2009089129 A1 WO 2009089129A1 US 2009030068 W US2009030068 W US 2009030068W WO 2009089129 A1 WO2009089129 A1 WO 2009089129A1
Authority
WO
WIPO (PCT)
Prior art keywords
frames
camera
data
dimensional
image data
Prior art date
Application number
PCT/US2009/030068
Other languages
French (fr)
Inventor
Hongsheng Zhang
Original Assignee
3M Innovative Properties Company
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 3M Innovative Properties Company filed Critical 3M Innovative Properties Company
Priority to US12/811,239 priority Critical patent/US8803958B2/en
Priority to DE112009000101T priority patent/DE112009000101T5/en
Publication of WO2009089129A1 publication Critical patent/WO2009089129A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • G06V20/653Three-dimensional objects by matching three-dimensional models, e.g. conformal mapping of Riemann surfaces
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61CDENTISTRY; APPARATUS OR METHODS FOR ORAL OR DENTAL HYGIENE
    • A61C13/00Dental prostheses; Making same
    • A61C13/0003Making bridge-work, inlays, implants or the like
    • A61C13/0004Computer-assisted sizing or machining of dental prostheses
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/003Navigation within 3D models or images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/207Image signal generators using stereoscopic image cameras using a single 2D image sensor
    • H04N13/221Image signal generators using stereoscopic image cameras using a single 2D image sensor using the relative movement between cameras and objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/41Medical

Definitions

  • a method of three-dimensional reconstruction includes acquiring a plurality of frames of image data of a surface of an object, captured from a camera position along a camera path and including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object; selecting a subset of the frames of image data to provide a plurality of key frames, each one related to at least one other key frame by a portion of the camera path including a rotation and a translation determined based upon one or more common points in the three-dimensional reconstruction, wherein the remaining plurality of frames of image data are non-key frames; providing a three- dimensional model of the object; determining a second rotation and a second translation from one of the key frames to a non-key frame sequentially positioned between the one of the key frames and a sequentially adjacent key frame; obtaining three-dimensional reconstruction information of the surface of the object from the camera position at least one of the non-key frames to provide upsampled three-dimensional data; and
  • a method for interactively reducing accumulated error in a global path includes acquiring a plurality of frames of image - A -
  • the data for a three-dimensional reconstruction of the surface of the object may be obtained from at least one other channel image to provide disparity data. Acquiring one or more frames of image data along the recommended scan path may be accomplished to reduce the accumulated error. Two of the plurality of frames may be identified which further comprises identifying frames of image data that are separated by a substantially greater distance along the camera path than along the surface of the object.
  • a subset of the plurality of frames of image data may be selected to provide a plurality of key frames, each one of the plurality of key frames related to at least one other one of the plurality of key frames by a portion of the camera path including a rotation and a translation determined based upon one or more common points in the three-dimensional reconstruction of the surface of the object in each of the respective key frames, wherein the remaining ones of the plurality of frames of image data are non-key frames.
  • the weighting matrix may be selected to locally decouple the error function around a centroid of common surface data for two or more three- dimensional reconstructions. An error function may be minimized to evaluate a calibration state based on the resulting error function minimization.
  • image generally refers to a two- dimensional set of pixels forming a two-dimensional view of a subject within an image plane.
  • image set generally refers to a set of related two-dimensional images that might be resolved into three-dimensional data.
  • point cloud generally refers to a three-dimensional set of points forming a three-dimensional view of the subject reconstructed from a number of two-dimensional images. In a three-dimensional image capture system, a number of such point clouds may also be registered and combined into an aggregate point cloud constructed from images captured by a moving camera.
  • pixels generally refer to two-dimensional data and points generally refer to three-dimensional data, unless another meaning is specifically indicated or clear from the context.
  • a tool or control may also include any physical hardware relating to the user input, such as a mouse, keyboard, display, keypad, track ball, and/or any other device that receives physical input from a user and converts the physical input into an input for use in a computerized system.
  • any physical hardware relating to the user input such as a mouse, keyboard, display, keypad, track ball, and/or any other device that receives physical input from a user and converts the physical input into an input for use in a computerized system.
  • the camera 102 is a handheld, freely-positionable probe having at least one user-input device 116, such as a button, a lever, a dial, a thumb wheel, a switch, or the like, for user control of the image capture system 100 such as starting and stopping scans.
  • the camera 102 may be shaped and sized for dental scanning. More particularly, the camera 102 may be shaped and sized for intraoral scanning and data capture, such as by insertion into a mouth of an imaging subject and passing over an intraoral surface 106 at a suitable distance to acquire surface data from teeth, gums, and so forth.
  • real time more specifically refers to processing within the time between frames of video data, which may vary according to specific video technologies between about fifteen frames per second and about thirty frames per second.
  • processing capabilities of the computer 108 may vary according to the size of the object 104, the speed of image acquisition, and the desired spatial resolution of three-dimensional points.
  • the computer 108 may also include peripheral devices such as a keyboard 114, display 110, and mouse 112 for user interaction with the camera system 100.
  • the display 110 may be a touch screen display capable of receiving user input through direct, physical interaction with the display 110.
  • the display may include an autostereoscopic display or the like capable of displaying stereo images.
  • the camera 102 may acquire two-dimensional image sets at a video rate while the camera 102 is passed over a surface of the subject.
  • the two-dimensional image sets may be forwarded to the computer 108 for derivation of three-dimensional point clouds.
  • the three-dimensional data for each newly acquired two-dimensional image set may be derived and fitted or "stitched" to existing three-dimensional data using a number of different techniques.
  • Such a system may employ camera motion estimation to avoid the need for independent tracking of the position of the camera 102.
  • One useful example of such a technique is described in commonly-owned U.S. App. No. 11/270,135, filed on November 9, 2005, the entire content of which is incorporated herein by reference. However, it will be appreciated that this example is not limiting, and that the principles described herein may be applied to a wide range of three-dimensional image capture systems.
  • the system 100 may include a computer-usable or computer-readable medium.
  • the computer-usable medium 118 may include one or more memory chips (or other chips, such as a processor, that include memory), optical disks, magnetic disks or other magnetic media, and so forth.
  • the computer-usable medium 118 may in various embodiments include removable memory (such as a USB device, tape drive, external hard drive, and so forth), remote storage (such as network attached storage), volatile or non- volatile computer memory, and so forth.
  • the computer-usable medium 118 may contain computer-readable instructions for execution by the computer 108 to perform the various processes described herein.
  • the computer-usable medium 118 may also, or instead, store data received from the camera 102, store a three-dimensional model of the object 104, store computer code for rendering and display, and so forth.
  • This center channel image may be presented in a user interface to permit inspection, marking, and other manipulation by a user during a user session as describe below.
  • Fig. 6A illustrates a camera path in a world coordinate system.
  • the camera begins at a starting point 610 and follows a path 620 in a counterclockwise direction as indicated by an arrow 625, returning to an ending point coincident with the starting point 610 in a fixed coordinate system, such as an arbitrarily selected world coordinate system.
  • Fig. 6B shows a camera path in a camera coordinate system.
  • errors may accumulate in a calculated camera path 635 so that a measured ending point 640 appears to be located away from the measured starting point 630 in the camera coordinate system, even though these points are identical in the world coordinate system.
  • one or more cross links such as those described above with reference to Fig. 4 may be employed to mitigate accumulated errors in the calculated camera path 635.
  • the process 700 may begin with preprocessing as shown in step 710. It will be understood that preprocessing as described herein presupposes the availability of a number of frames of image data from which a camera path and three-dimensional model can be reconstructed.
  • the information for the three-dimensional reconstruction may be generated in numerous ways including coming from structured light projection, shading based three-dimensional reconstruction, or disparity data. Disparity data may be generated by a conventional image plus one or more other channels or side channels.
  • the preprocessing may include determining the number of available frames, the amount of overlap between neighboring frames, identification and elimination of frames with blurred or badly distorted images, and any other suitable preprocessing steps. An estimate of the number of desired key frames may be initially determined during the preprocessing step.
  • key frames may be selected from among all of the frames of data acquired from a camera along a camera path.
  • computational costs can be reduced by storing certain data and performing certain calculations and processing steps exclusively with reference to key frames.
  • These key frames may be related to one another in a manner that permits characterization of a complete camera path, typically through the registration of overlapping three-dimensional data in respective key frames.
  • Various methods are known in the art for selecting a subset of frames of data as key frames, including techniques based on image overlap, camera path distance, the number of intervening non-key frames and so forth.
  • Key frames may also or instead be selected based upon an amount of image overlap from the preceding key frame and/or a candidate for a following key frame (if available).
  • a graph analysis may be performed using the key frames and the associated stitching to calculate a global path for the camera used to obtain a three-dimensional model.
  • the graph analysis may consider each key frame as a node or vertex and each stitch as an edge between a pair of nodes.
  • a key frame is selected as a starting point.
  • a breadth- or depth-first search may be performed through the graph to identify stitches which may connect the current key frame to another key frame.
  • Each key frame may be marked as the graph is processed.
  • a check may be performed to see if all key frames have been reached within the graph. If all key frames have not been reached through traversing stitches in the graph analysis, the largest subgraph is identified. This sub-graph may be examined to see if the entire three- dimensional image may be modeled.
  • step 712 may be performed. For example, a set of key frames may have been selected which did not have sufficient stitching from one key frame to the next key frame. By choosing a different set of key frames, sufficient stitching may be obtained in order to obtain a complete graph of all needed aspects of the three-dimensional imaging.
  • a numerical optimization may be performed to reduce errors in the calculated camera path based upon available data for the complete camera path such as, for example, cross links that interrelate temporally distant measurements.
  • the objective of numerical optimization is to minimize a calculated error based upon an error function for the camera path and/or reconstructed three-dimensional model.
  • the representation of a point in one camera's coordinate system may be related to the same point in another coordinate system.
  • coordinates of a point, X may be transformed from the A coordinate system to the B coordinate system as follows:
  • the motion constraints may form an overdetermined system of motion constraint equations. Using these equations as a starting point, numerical optimization may be performed on the rotational and translational components of each camera based on the measured stitches.
  • the rotational and translational components may be independently optimized.
  • R c ' the corresponding candidate camera-to-camera rotations, R c ' AB
  • R c ' AB the corresponding candidate camera-to-camera rotations
  • R AB the corresponding residual rotations
  • R r ⁇ es ⁇ dual AB R c ' AB (R A ' B ) ⁇ 1 •
  • a scalar- valued rotational cost function, e r may be computed that depends on the candidate camera rotations
  • the rotational error function may be converted into a quaternion expression in order to translate the numerical problem into a linear system of equations for solution. While this approach may increase computational efficiency, it offers an incomplete optimization solution.
  • upsampling may be performed to augment a three- dimensional model with data from non-key frames.
  • non-key frames may be registered to nearby key frames to create small, local reconstruction patches including the full image detail available from non-key frames.
  • path optimization may be performed on a key- frame-based data set, thus reducing the data requiring processing, while retaining additional data points from non-key frames for use in the final three- dimensional model.
  • a realization may include computer executable code created using a structured programming language such as C, an object oriented programming language such as C++, or any other high-level or low-level programming language (including assembly languages, hardware description languages, and database programming languages and technologies) that may be stored, compiled or interpreted to run on one of the above devices, as well as heterogeneous combinations of processors, processor architectures, or combinations of different hardware and software.
  • a computer program product comprising computer executable code that, when executing on one or more computing devices, performs any and/or all of the steps described above.

Abstract

Disclosed herein are various techniques for improving global path optimization in a system that uses camera path for three-dimensional reconstruction. A subset of frames of data for the global path, the key frames, may be used to reduce the computational complexity of the optimization, while preserving full three-dimensional detail in the optimized model by relating other measurements to the optimized key frame path.

Description

GLOBAL CAMERA PATH OPTIMIZATION
[0001] The present application claims priority from the U.S. Provisional Patent Application No. 61/019,159, filed January 4, 2008, which is hereby incorporated by reference in its entirety.
FIELD OF INVENTION
[0002] This invention relates generally to three-dimensional imaging and more specifically to optimizing the calculation of a global camera path used in a three- dimensional reconstruction.
BACKGROUND
[0003] In one technique for three-dimensional image reconstruction, a number of images or image sets of an object are captured with a camera that travels in a path over the surface of the object. Information from this catalogue of information can then be used to reconstruct a three-dimensional model of the object based upon the camera path and individual three-dimensional measurements captured along the camera path. The path of a camera may be very long and complex involving motion estimation from image to image that accumulates significant errors along its length. These errors can result in a variety of reconstruction artifacts in a resulting three-dimensional model such as double surfaces where the camera path scans the same region twice with an error in camera position between the two scans. Error may also be introduced due to calibration problems, inaccuracies in camera distortion models used to determine three-dimensional data, and so forth.
[0004] While various techniques exist for minimizing errors along an entire camera path, there remains a need for improved global path optimization techniques suitable for use with data-intensive path optimizations typical of high-accuracy, three- dimensional reconstruction. SUMMARY
[0005] Disclosed herein are various techniques for improving global path optimization in a system that uses camera path for three-dimensional reconstruction. A subset of frames of data for the global path, the key frames, may be used to reduce the computational complexity of the optimization, while preserving full three-dimensional detail in the optimized model by relating other measurements to the optimized key frame path.
[0006] In one aspect, a method of three-dimensional reconstruction that is disclosed herein includes acquiring a plurality of frames of image data of a surface of an object, captured from a camera position along a camera path and including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object; selecting a subset of the frames of image data to provide a plurality of key frames, each one related to at least one other key frame by a portion of the camera path including a rotation and a translation determined based upon one or more common points in the three-dimensional reconstruction, wherein the remaining plurality of frames of image data are non-key frames; providing a three- dimensional model of the object; determining a second rotation and a second translation from one of the key frames to a non-key frame sequentially positioned between the one of the key frames and a sequentially adjacent key frame; obtaining three-dimensional reconstruction information of the surface of the object from the camera position at least one of the non-key frames to provide upsampled three-dimensional data; and adding the upsampled three-dimensional data to the three-dimensional model based upon the second rotation and the second translation. A camera motion may be estimated between two adjacent key frames based on the rotation and the translation. There may be optimizing the estimation of the camera motion between the two adjacent key frames by creating consistency among motion parameters using an overdetermined system of motion constraint equations, wherein the motion parameters are comprised of information on the rotation and the translation. There may also be optimizing the camera motion between two adjacent non-key frames by creating consistency among motion parameters using an overdetermined system of motion constraint equations. The three-dimensional reconstruction may be updated based on the created consistency among the motion parameters. The data for a three-dimensional reconstruction of the surface of the object may be obtained from at least one other channel image to provide disparity data. The three-dimensional model may further comprise generating a three-dimensional model of the object using the camera path and the three-dimensional reconstruction for each of the key frames. Three-dimensional reconstruction information may be obtained of the surface of the object from the camera position for all of the non-key frames between two adjacent key frames. A subset of the plurality of frames may be selected based on a quality metric of the three-dimensional reconstruction. Selecting of the subset of the plurality of frames may be determined using graph analysis to ensure that all of the key frames are utilized in the three-dimensional reconstruction.
[0007] In one aspect, a computer program product that is disclosed herein performs the steps of acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three- dimensional reconstruction of the surface of the object as viewed from the camera position; selecting a subset of the plurality of frames of image data to provide a plurality of key frames, each one of the plurality of key frames related to at least one other one of the plurality of key frames by a portion of the camera path including a rotation and a translation determined based upon one or more common points in the three-dimensional reconstruction of the surface of the object in each of the respective key frames, wherein the remaining ones of the plurality of frames of image data are non-key frames; providing a three-dimensional model of the object; determining a second rotation and a second translation from one of the key frames to at least one of the non-key frames sequentially positioned between the one of the key frames and a sequentially adjacent one of the key frames; obtaining three-dimensional reconstruction information of the surface of the object from the camera position of the at least one of the non-key frames to provide upsampled three-dimensional data; and adding the upsampled three-dimensional data to the three-dimensional model based upon the second rotation and the second translation.
[0008] In one aspect, a method for interactively reducing accumulated error in a global path is disclosed herein which includes acquiring a plurality of frames of image - A -
data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; generating a three-dimensional model of the object using the camera path and the data for the three-dimensional reconstruction; identifying two of the plurality of frames of image data that represent a candidate for an accumulated error in the camera path relative to one another; and displaying the three-dimensional model along with a graphical annotation that illustrates a recommended scan path to reduce the accumulated error. The data for a three-dimensional reconstruction of the surface of the object may be obtained from at least one other channel image to provide disparity data. Acquiring one or more frames of image data along the recommended scan path may be accomplished to reduce the accumulated error. Two of the plurality of frames may be identified which further comprises identifying frames of image data that are separated by a substantially greater distance along the camera path than along the surface of the object.
[0009] In one aspect, a computer program product comprising computer executable code embodied in a computer readable medium is disclosed herein which performs the steps of acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three- dimensional reconstruction of the surface of the object as viewed from the camera position; generating a three-dimensional model of the object using the camera path and the data for the three-dimensional reconstruction; identifying two of the plurality of frames of image data that represent a candidate for an accumulated error in the camera path relative to one another; and displaying the three-dimensional model along with a graphical annotation that illustrates a recommended scan path to reduce the accumulated error.
[0010] In one aspect, a system is disclosed herein which comprises a camera, a monitor, a processor, and a memory, the memory storing a computer program executable by the processor to perform the steps of acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; generating a three-dimensional model of the object using the camera path and the data for the three-dimensional reconstruction; identifying two of the plurality of frames of image data that represent a candidate for an accumulated error in the camera path relative to one another; and displaying the three-dimensional model along with a graphical annotation that illustrates a recommended scan path to reduce the accumulated error.
[0011] In one aspect, a method for global path optimization is disclosed herein includes acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; and minimizing an error function for a plurality of camera positions along the camera path, the error function including a system of equations for translational components of an error and for rotational components of the error, wherein the error function couples the translational components and the rotational components using a weighting matrix, thereby providing an optimized camera path. The data for a three-dimensional reconstruction of the surface of the object may be obtained from at least one other channel image to provide disparity data. The system of equations may be a non-linear system of equations. The translational component of the error may form of a system of linear equations. The rotational component of the error may form a system of non- linear equations. A three- dimensional model may be generated based upon the camera path and the data for the three-dimensional reconstruction, and the three-dimensional model may be refined based upon the optimized camera path. A subset of the plurality of frames of image data may be selected to provide a plurality of key frames, each one of the plurality of key frames related to at least one other one of the plurality of key frames by a portion of the camera path including a rotation and a translation determined based upon one or more common points in the three-dimensional reconstruction of the surface of the object in each of the respective key frames, wherein the remaining ones of the plurality of frames of image data are non-key frames. The weighting matrix may be selected to locally decouple the error function around a centroid of common surface data for two or more three- dimensional reconstructions. An error function may be minimized to evaluate a calibration state based on the resulting error function minimization.
[0012] In one aspect, a computer program product comprising computer executable code is disclosed herein performs the steps of acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; and minimizing an error function for a plurality of camera positions along the camera path, the error function including a system of equations for translational components of an error and for rotational components of the error, wherein the error function couples the translational components and the rotational components using a weighting matrix, thereby providing an optimized camera path.
BRIEF DESCRIPTION OF THE DRAWINGS
[0013] The invention and the following detailed description of certain embodiments thereof may be understood by reference to the following figures.
[0014] Fig. 1 shows a three-dimensional scanning system.
[0015] Fig. 2 shows a schematic diagram of an optical system for a three- dimensional camera.
[0016] Fig. 3 shows a processing pipeline for obtaining three-dimensional data from a video camera.
[0017] Fig. 4A and 4B illustrate camera paths for a three-dimensional camera.
[0018] Fig. 5 shows a user interface image where additional data is requested by a software tool.
[0019] Fig. 6A and 6B illustrate accumulated error in camera paths. [0020] Fig. 7 is a flow chart of a three-dimensional reconstruction process including global path optimization for improved accuracy.
DETAILED DESCRIPTION
[0021] In the following text, references to items in the singular should be understood to include items in the plural, and vice versa, unless explicitly stated otherwise or clear from the text. Grammatical conjunctions are intended to express any and all disjunctive and conjunctive combinations of conjoined clauses, sentences, words, and the like, unless otherwise stated or clear from the context.
[0022] In the systems and methods described herein, a number of techniques for global motion optimization are employed to improve accuracy of three-dimensional reconstructions based upon camera path.
[0023] The following description details specific scanning technologies and focuses on dental applications of three-dimensional imaging; however, it will be appreciated that variations, adaptations, and combinations of the methods and systems below will be apparent to one of ordinary skill in the art. For example, while an image- based system is described, non-image based scanning techniques such as infrared time- of-flight techniques or structured light techniques using patterned projections may similarly employ reconstruction based on camera path that may benefit from the improvements described herein. As another example, while digital dentistry is one useful application of the improved accuracy that results from the techniques described herein, global path optimization may also usefully be employed to refine three-dimensional animation models or three-dimensional scans for machine vision applications or for mapping applications. All such variations, adaptations, and combinations are intended to fall within the scope of this disclosure.
[0024] In the following description, the term "image" generally refers to a two- dimensional set of pixels forming a two-dimensional view of a subject within an image plane. The term "image set" generally refers to a set of related two-dimensional images that might be resolved into three-dimensional data. The term "point cloud" generally refers to a three-dimensional set of points forming a three-dimensional view of the subject reconstructed from a number of two-dimensional images. In a three-dimensional image capture system, a number of such point clouds may also be registered and combined into an aggregate point cloud constructed from images captured by a moving camera. Thus it will be understood that pixels generally refer to two-dimensional data and points generally refer to three-dimensional data, unless another meaning is specifically indicated or clear from the context.
[0025] The terms "three-dimensional model", "three-dimensional surface representation", "digital surface representation", "three-dimensional surface map", and the like, as used herein, are intended to refer to any three-dimensional surface map of an object, such as a point cloud of surface data, a set of two-dimensional polygons, or any other data representing all or some of the surface of an object, as might be obtained through the capture and/or processing of three-dimensional scan data, unless a different meaning is explicitly provided or otherwise clear from the context. A "three-dimensional representation" may include any of the three-dimensional surface representations described above, as well as volumetric and other representations, unless a different meaning is explicitly provided or otherwise clear from the context.
[0026] In general, the terms "render" or "rendering" refer to a two-dimensional visualization of a three-dimensional object, such as for display on a monitor. However, it will be understood that a variety of three-dimensional rendering technologies exist, and may be usefully employed with the systems and methods disclosed herein. For example, the system and methods described herein may usefully employ a holographic display, an autostereoscopic display, an anaglyph display, a head-mounted stereo display, or any other two-dimensional and/or three-dimensional display. As such, rendering as described herein should be interpreted broadly unless a narrower meaning is explicitly provided or otherwise clear from the context.
[0027] The term "dental object", as used herein, is intended to refer broadly to subject matter related to dentistry. This may include intraoral structures such as dentition, and more typically human dentition, such as individual teeth, quadrants, full arches, pairs of arches (which may be separate or in occlusion of various types), soft tissue, and the like, as well bones and any other supporting or surrounding structures. As used herein, the term "intraoral structures" refers to both natural structures within a mouth as described above and artificial structures such as any of the dental objects described below that might be present in the mouth. Dental objects may include "restorations", which may be generally understood to include components that restore the structure or function of existing dentition, such as crowns, bridges, veneers, inlays, onlays, amalgams, composites, and various substructures such as copings and the like, as well as temporary restorations for use while a permanent restoration is being fabricated. Dental objects may also include a "prosthesis" that replaces dentition with removable or permanent structures, such as dentures, partial dentures, implants, retained dentures, and the like. Dental objects may also include "appliances" used to correct, align, or otherwise temporarily or permanently adjust dentition, such as removable orthodontic appliances, surgical stents, bruxism appliances, snore guards, indirect bracket placement appliances, and the like. Dental objects may also include "hardware" affixed to dentition for an extended period, such as implant fixtures, implant abutments, orthodontic brackets, and other orthodontic components. Dental objects may also include "interim components" of dental manufacture such as dental models (full and/or partial), wax-ups, investment molds, and the like, as well as trays, bases, dies, and other components employed in the fabrication of restorations, prostheses, and the like. Dental objects may also be categorized as natural dental objects such as the teeth, bone, and other intraoral structures described above or as artificial dental objects such as the restorations, prostheses, appliances, hardware, and interim components of dental manufacture as described above.
[0028] Terms such as "digital dental model", "digital dental impression" and the like, are intended to refer to three-dimensional representations of dental objects that may be used in various aspects of acquisition, analysis, prescription, and manufacture, unless a different meaning is otherwise provided or clear from the context. Terms such as "dental model" or "dental impression" are intended to refer to a physical model, such as a cast, printed, or otherwise fabricated physical instance of a dental object. Unless specified, the term "model", when used alone, may refer to either or both of a physical model and a digital model.
[0029] It will further be understood that terms such as "tool" or "control", when used to describe aspects of a user interface, are intended to refer generally to a variety of techniques that may be employed within a graphical user interface or other user interface to receive user input that stimulates or controls processing including without limitation drop-down lists, radio buttons, cursor and/or mouse actions (selections by point, selections by area, drag-and-drop operations, and so forth), check boxes, command lines, text input fields, messages and alerts, progress bars, and so forth. A tool or control may also include any physical hardware relating to the user input, such as a mouse, keyboard, display, keypad, track ball, and/or any other device that receives physical input from a user and converts the physical input into an input for use in a computerized system. Thus in the following description the terms "tool", "control" and the like should be broadly construed unless a more specific meaning is otherwise provided or clear from the context.
[0030] Fig. 1 depicts a three-dimensional scanning system that may be used with the systems and methods described herein. In general, the system 100 may include a camera 102 that captures images from a surface 106 of an object 104, such as a dental patient, and forwards the images to a computer 108, which may include a display 110 and one or more user-input devices 112, 114 such as a mouse 112 or a keyboard 114. The camera 102 may also include an integrated input or output device 116 such as a control input (e.g., button, touchpad, thumbwheel, etc.) or a display (e.g., LCD or LED display) to provide status information.
[0031] The camera 102 may include any camera or camera system suitable for capturing images from which a three-dimensional point cloud or other three-dimensional data may be recovered. For example, the camera 102 may employ a multi-aperture system as disclosed in U.S. Pat. No. 7,372,642 to Rohaly et al., the entire content of which is incorporated herein by reference. While Rohaly discloses one multi-aperture system, it will be appreciated that any multi-aperture system suitable for reconstructing a three-dimensional point cloud from a number of two-dimensional images may similarly be employed. In one multi-aperture embodiment, the camera 102 may include a plurality of apertures including a center aperture positioned along a center optical axis of a lens that provides a center channel for the camera 102, along with any associated imaging hardware. In such embodiments, the center channel may provide a conventional video image of the scanned subject matter, while a number of axially offset channels yield image sets containing disparity information that can be employed in three-dimensional reconstruction of a surface. In other embodiments, a separate video camera and/or channel may be provided to achieve the same result, i.e., a video of an object corresponding temporally to a three-dimensional scan of the object, preferably from the same perspective, or from a perspective having a fixed, known relationship to the perspective of the camera 102. The camera 102 may also, or instead, include a stereoscopic, triscopic or other multi-camera or other configuration in which a number of cameras or optical paths are maintained in fixed relation to one another to obtain two- dimensional images of an object from a number of different perspectives. The camera 102 may include suitable processing for deriving a three-dimensional point cloud from an image set or a number of image sets, or each two-dimensional image set may be transmitted to an external processor such as contained in the computer 108 described below. In other embodiments, the camera 102 may employ structured light, laser scanning, direct ranging, or any other technology suitable for acquiring three-dimensional data, or two-dimensional data that can be resolved into three-dimensional data. While the techniques described below can usefully employ video data acquired by a video-based three-dimensional scanning system, it will be understood that any other three- dimensional scanning system may be supplemented with a video acquisition system that captures suitable video data contemporaneously with, or otherwise synchronized with, the acquisition of three-dimensional data.
[0032] In one embodiment, the camera 102 is a handheld, freely-positionable probe having at least one user-input device 116, such as a button, a lever, a dial, a thumb wheel, a switch, or the like, for user control of the image capture system 100 such as starting and stopping scans. In an embodiment, the camera 102 may be shaped and sized for dental scanning. More particularly, the camera 102 may be shaped and sized for intraoral scanning and data capture, such as by insertion into a mouth of an imaging subject and passing over an intraoral surface 106 at a suitable distance to acquire surface data from teeth, gums, and so forth. The camera 102 may, through such a continuous data acquisition process, capture a point cloud of surface data having sufficient spatial resolution and accuracy to prepare dental objects such as prosthetics, hardware, appliances, and the like therefrom, either directly or through a variety of intermediate processing steps. In other embodiments, surface data may be acquired from a dental model such as a dental prosthesis, to ensure proper fitting using a previous scan of corresponding dentition, such as a tooth surface prepared for the prosthesis. [0033] Although not shown in Fig. 1 , it will be appreciated that a number of supplemental lighting systems may be usefully employed during image capture. For example, environmental illumination may be enhanced with one or more spotlights illuminating the object 104 to speed image acquisition and improve depth of field (or spatial resolution depth). The camera 102 may also, or instead, include a strobe, a flash, or some other light source to supplement illumination of the object 104 during image acquisition.
[0034] The object 104 may be any object, collection of objects, portion of an object, or other subject matter. More particularly with respect to the dental techniques discussed herein, the object 104 may include human dentition captured intraorally from a dental patient's mouth. A scan may capture a three-dimensional representation of some or all of the dentition according to a particular purpose of the scan. Thus the scan may capture a digital model of a tooth, a quadrant of teeth, or a full collection of teeth including two opposing arches, as well as soft tissue or any other relevant intraoral structures. The scan may capture multiple representations, such as a tooth surface before and after preparation for a restoration. As will be noted below, this data may be employed for subsequent modeling such as designing a restoration or determining a margin line for same. During the scan, a center channel of the camera 102 or a separate video system may capture video of the dentition from the point of view of the camera 102. In other embodiments where, for example, a completed fabrication is being virtually test fitted to a surface preparation, the scan may include a dental prosthesis such as an inlay, a crown, or any other dental prosthesis, dental hardware, dental appliance, or the like. The object 104 may also, or instead, include a dental model, such as a plaster cast, a wax-up, an impression, or a negative impression of a tooth, teeth, soft tissue, or some combination of these.
[0035] The computer 108 may include, for example, a personal computer or other processing device. In one embodiment, the computer 108 includes a personal computer with a dual 2.8 GHz Opteron central processing unit, 2 gigabytes of random access memory, a TYAN Thunder K8WE motherboard, and a 250 gigabyte, 10,000 rpm hard drive. In one current embodiment, the system can be operated to capture more than five thousand points per image set in real time using the techniques described herein, and store an aggregated point cloud of several million points. Of course, this point cloud may be further processed to accommodate subsequent data handling, such as by decimating the point cloud data or generating a corresponding mesh of surface data. As used herein, the term "real time" means generally with no observable latency between processing and display. In a video-based scanning system, real time more specifically refers to processing within the time between frames of video data, which may vary according to specific video technologies between about fifteen frames per second and about thirty frames per second. More generally, processing capabilities of the computer 108 may vary according to the size of the object 104, the speed of image acquisition, and the desired spatial resolution of three-dimensional points. The computer 108 may also include peripheral devices such as a keyboard 114, display 110, and mouse 112 for user interaction with the camera system 100. The display 110 may be a touch screen display capable of receiving user input through direct, physical interaction with the display 110. In another aspect, the display may include an autostereoscopic display or the like capable of displaying stereo images.
[0036] Communications between the computer 108 and the camera 102 may use any suitable communications link including, for example, a wired connection or a wireless connection based upon, for example, IEEE 802.11 (also known as wireless Ethernet), Blue Tooth, or any other suitable wireless standard using, e.g., a radio frequency, infrared, or other wireless communication medium. In medical imaging or other sensitive applications, wireless image transmission from the camera 102 to the computer 108 may be secured. The computer 108 may generate control signals to the camera 102 which, in addition to image acquisition commands, may include conventional camera controls such as focus or zoom.
[0037] In an example of general operation of a three-dimensional image capture system 100, the camera 102 may acquire two-dimensional image sets at a video rate while the camera 102 is passed over a surface of the subject. The two-dimensional image sets may be forwarded to the computer 108 for derivation of three-dimensional point clouds. The three-dimensional data for each newly acquired two-dimensional image set may be derived and fitted or "stitched" to existing three-dimensional data using a number of different techniques. Such a system may employ camera motion estimation to avoid the need for independent tracking of the position of the camera 102. One useful example of such a technique is described in commonly-owned U.S. App. No. 11/270,135, filed on November 9, 2005, the entire content of which is incorporated herein by reference. However, it will be appreciated that this example is not limiting, and that the principles described herein may be applied to a wide range of three-dimensional image capture systems.
[0038] The display 110 may include any display suitable for video or other rate rendering at a level of detail corresponding to the acquired data. Suitable displays include cathode ray tube displays, liquid crystal displays, light emitting diode displays and the like. In general, the display 110 may be operative Iy coupled to, and capable of receiving display signals from, the computer 108. This display may include a CRT or flat panel monitor, a three-dimensional display (such as an anaglyph display), an autostereoscopic three-dimensional display or any other suitable two-dimensional or three-dimensional rendering hardware. In some embodiments, the display may include a touch screen interface using, for example capacitive, resistive, or surface acoustic wave (also referred to as dispersive signal) touch screen technologies, or any other suitable technology for sensing physical interaction with the display 110.
[0039] The system 100 may include a computer-usable or computer-readable medium. The computer-usable medium 118 may include one or more memory chips (or other chips, such as a processor, that include memory), optical disks, magnetic disks or other magnetic media, and so forth. The computer-usable medium 118 may in various embodiments include removable memory (such as a USB device, tape drive, external hard drive, and so forth), remote storage (such as network attached storage), volatile or non- volatile computer memory, and so forth. The computer-usable medium 118 may contain computer-readable instructions for execution by the computer 108 to perform the various processes described herein. The computer-usable medium 118 may also, or instead, store data received from the camera 102, store a three-dimensional model of the object 104, store computer code for rendering and display, and so forth.
[0040] Fig. 2 depicts an optical system 200 for a three-dimensional camera that may be used with the systems and methods described herein, such as for the camera 102 described above with reference to Fig. 1. [0041] The optical system 200 may include a primary optical facility 202, which may be employed in any kind of image processing system. In general, a primary optical facility refers herein to an optical system having one optical channel. Typically, this optical channel shares at least one lens, and has a shared image plane within the optical system, although in the following description, variations to this may be explicitly described or otherwise clear from the context. The optical system 200 may include a single primary lens, a group of lenses, an object lens, mirror systems (including traditional mirrors, digital mirror systems, digital light processors, or the like), confocal mirrors, and any other optical facilities suitable for use with the systems described herein. The optical system 200 may be used, for example in a stereoscopic or other multiple image camera system. Other optical facilities may include holographic optical elements or the like. In various configurations, the primary optical facility 202 may include one or more lenses, such as an object lens (or group of lenses) 202b, a field lens 202d, a relay lens 202f, and so forth. The object lens 202b may be located at or near an entrance pupil 202a of the optical system 200. The field lens 202d may be located at or near a first image plane 202c of the optical system 200. The relay lens 202f may relay bundles of light rays within the optical system 200. The optical system 200 may further include components such as aperture elements 208 with one or more apertures 212, a refocusing facility 210 with one or more refocusing elements 204, one or more sampling facilities 218, and/or a number of sensors 214a, 214b, 214c.
[0042] The optical system 200 may be designed for active wavefront sampling, which should be understood to encompass any technique used to sample a series or collection of optical data from an object 220 or objects, including optical data used to help detect two-dimensional or three-dimensional characteristics of the object 220, using optical data to detect motion, using optical data for velocimetry or object tracking, or the like. Further details of an optical system that may be employed as the optical system 200 of Fig. 2 are provided in U.S. Pat. No. 7,372,642, the entire content of which is incorporated herein by reference. More generally, it will be understood that, while Fig. 2 depicts one embodiment of an optical system 200, numerous variations are possible. One salient feature of the optical system related to the discussion below is the use of a center optical channel that captures conventional video or still images at one of the sensors 214b concurrent with various offset data (at, e.g., 214a and 214c) used to capture three- dimensional measurements. This center channel image may be presented in a user interface to permit inspection, marking, and other manipulation by a user during a user session as describe below.
[0043] Fig. 3 shows a three-dimensional reconstruction system 300 employing a high-speed pipeline and a high-accuracy pipeline. In general, the high-speed processing pipeline 330 aims to provide three-dimensional data in real time, such as at a video frame rate used by an associated display, while the high-accuracy processing pipeline 350 aims to provide the highest accuracy possible from camera measurements, subject to any external computation or time constraints imposed by system hardware or an intended use of the results. A data source 310 such as the camera 102 described above provides image data or the like to the system 300. The data source 310 may for example include hardware such as LED ring lights, wand sensors, a frame grabber, a computer, an operating system and any other suitable hardware and/or software for obtaining data used in a three-dimensional reconstruction. Images from the data source 310, such as center channel images containing conventional video images and side channels containing disparity data used to recover depth information may be passed to the real-time processing controller 316. The real-time processing controller 316 may also provide camera control information or other feedback to the data source 310 to be used in subsequent data acquisition or for specifying data already obtained in the data source 310 that is needed by the real-time processing controller 316. Full resolution images and related image data may be retained in a full resolution image store 322. The stored images may, for example, be provided to the high-accuracy processing controller 324 during processing, or be retained for image review by a human user during subsequent processing steps.
[0044] The real-time processing controller 316 may provide images or frames to the high-speed (video rate) processing pipeline 330 for reconstruction of three- dimensional surfaces from the two-dimensional source data in real time. In an exemplary embodiment, two-dimensional images from an image set such as side channel images, may be registered by a two-dimensional image registration module 332. Based on the results of the two-dimensional image registration, a three-dimensional point cloud generation module 334 may create a three-dimensional point cloud or other three- dimensional representation. The three-dimensional point clouds from individual image sets may be combined by a three-dimensional stitching module 336. Finally, the stitched measurements may be combined into an integrated three-dimensional model by a three- dimensional model creation module 338. The resulting model may be stored as a highspeed three-dimensional model 340.
[0045] The high-accuracy processing controller 324 may provide images or frames to the high-accuracy processing pipeline 350. Separate image sets may have two- dimensional image registration performed by a two-dimensional image registration module 352. Based on the results of the two-dimensional image registration a three- dimensional point cloud or other three-dimensional representation may be generated by a three-dimensional point cloud generation module 354. The three-dimensional point clouds from individual image sets may be connected using a three-dimensional stitching module 356. Global motion optimization, also referred to herein as global path optimization or global camera path optimization, may be performed by a global motion optimization module 357 in order to reduce errors in the resulting three-dimensional model 358. In general, the path of the camera as it obtains the image frames may be calculated as a part of the three-dimensional reconstruction process. In a post-processing refinement procedure, the calculation of camera path may be optimized - that is, the accumulation of errors along the length of the camera path may be minimized by supplemental frame-to-frame motion estimation with some or all of the global path information. Based on global information such as individual frames of data in the image store 322, the high-speed three-dimensional model 340, and intermediate results in the high-accuracy processing pipeline 350, the high-accuracy model 370 may be processed to reduce errors in the camera path and resulting artifacts in the reconstructed model. As a further refinement, a mesh may be projected onto the high-speed model by a mesh projection module 360. The resulting images may be warped or deformed by a warping module 362. Warped images may be used to ease alignment and stitching between images, such as by reducing the initial error in a motion estimate. The warped images may be provided to the two-dimensional image registration module 352. The feedback of the high-accuracy three-dimensional model 370 into the pipeline may be repeated until some metric is obtained, such as a stitching accuracy or a minimum error threshold.
[0046] Various aspects of the system 300 of Fig. 3 are described in greater detail below. In particular, a model refinement process is described that may be used by the high-accuracy processing controller 324 to refine the high accuracy three-dimensional model 370 using measured data in the image store 322. It should be understood that various processing modules, or the steps implied by the modules, shown in this figure are exemplary in nature and that the order of processing, or the steps of the processing sequence, may be modified, omitted, repeated, re-ordered, or supplemented, without departing from the scope of this disclosure.
[0047] Fig. 4A shows an object 410 for imaging, along with a path 415 that a camera may follow while obtaining a three-dimensional scan of a surface of the object 410. The direction of the path 415 is indicated generally by an arrow 416. The object 410 may be an upper dental impression (as shown) or any other object from which three- dimensional surface data is sought. Starting the camera at a starting point 420, the camera may follow an arc 430 to a second point 422. The camera may then follow a segment 432 to a third point 424. The camera may then follow a second arc 434 to a fourth point 426. The camera may then follow a second segment 436 to return approximately to the starting point 420. It should be noted that the path 415 followed by the camera may be irregular rather than smooth, and that while a particular path 415 is depicted, more generally any path may be taken by the camera including paths that double back upon themselves, cross over identical regions two or more times, and/or entirely skip various surfaces of the object 410. It should also be noted that the camera path 415 may usefully return to the starting point 420, but this is not strictly required for three-dimensional reconstruction as described herein. The camera may take hundreds or thousands of images or more as the camera traverses the path around such a dental object 410.
[0048] Fig. 4B shows locations where additional scan data might usefully be acquired to improve the accuracy of a three-dimensional reconstruction. For example, arcs 440, 442, 444, and 446 may be scanned (e.g., traversed by the camera path) to provide cross linking between various lengths of the camera path. Data might usefully be acquired, for example, from any area that can improve computational accuracy of a three- dimensional reconstruction such as regions where the length of a camera path between two measurements of the surface (e.g., image sets or image data) is significantly greater than the distance between the two corresponding surface locations in the world coordinate system for the camera path. As another example, this may include regions where separate three-dimensional measurements for a general region of the reconstructed three-dimensional model fail to register to one another, or more generally where portions of the model or individual measurements contain indicia of accumulated error in the global camera path. Key frames (as described below) may be used to focus this inquiry on a subset of measurements that provide coverage for all or a substantial portion of the scanned subject matter.
[0049] Fig. 5 shows a user interface depicting a graphical request for additional scan data. After the camera follows the path 415 illustrated above, a software tool may be used to identify various locations where additional data might usefully be acquired to reduce accumulated error in a global camera path, as discussed above. A monitor 510 may display an image 520 such as a three-dimensional reconstruction of scanned subject matter, and an arrow 530 may be displayed on the monitor 510 indicating where additional scanning is recommended. The user may then proceed to use a camera, such as the camera 102 from Fig. 1, to scan the area indicated by the arrow 530. More generally, areas for additional scanning may be identified to a user in a graphical user interface that displays a reconstructed three-dimensional model from the camera path, along with arrows or other identifiers or graphical annotations that illustrate a recommended scan path. After a user augments a camera path with additional scans, the resulting data can be employed to resolve differences (i.e., errors) in the global camera path, as described generally throughout this disclosure.
[0050] Fig. 6A illustrates a camera path in a world coordinate system. The camera begins at a starting point 610 and follows a path 620 in a counterclockwise direction as indicated by an arrow 625, returning to an ending point coincident with the starting point 610 in a fixed coordinate system, such as an arbitrarily selected world coordinate system. [0051] Fig. 6B shows a camera path in a camera coordinate system. When a camera traverses the path 620 in the world coordinate system, errors may accumulate in a calculated camera path 635 so that a measured ending point 640 appears to be located away from the measured starting point 630 in the camera coordinate system, even though these points are identical in the world coordinate system. In one aspect, one or more cross links such as those described above with reference to Fig. 4 may be employed to mitigate accumulated errors in the calculated camera path 635.
[0052] Fig. 7 is a flow chart of a three-dimensional reconstruction process including global path optimization for improved accuracy.
[0053] The process 700 may begin with preprocessing as shown in step 710. It will be understood that preprocessing as described herein presupposes the availability of a number of frames of image data from which a camera path and three-dimensional model can be reconstructed. The information for the three-dimensional reconstruction may be generated in numerous ways including coming from structured light projection, shading based three-dimensional reconstruction, or disparity data. Disparity data may be generated by a conventional image plus one or more other channels or side channels. The preprocessing may include determining the number of available frames, the amount of overlap between neighboring frames, identification and elimination of frames with blurred or badly distorted images, and any other suitable preprocessing steps. An estimate of the number of desired key frames may be initially determined during the preprocessing step.
[0054] As shown in step 712, key frames may be selected from among all of the frames of data acquired from a camera along a camera path. In general, computational costs can be reduced by storing certain data and performing certain calculations and processing steps exclusively with reference to key frames. These key frames may be related to one another in a manner that permits characterization of a complete camera path, typically through the registration of overlapping three-dimensional data in respective key frames. Various methods are known in the art for selecting a subset of frames of data as key frames, including techniques based on image overlap, camera path distance, the number of intervening non-key frames and so forth. Key frames may also or instead be selected based upon an amount of image overlap from the preceding key frame and/or a candidate for a following key frame (if available). Too little overlap may impair frame -to-frame registration. Too much overlap may produce excess key frames requiring additional processing. Key frames may be selected based on spatial displacement. Key frames may also be selected based on sequential displacement. This type of sequential displacement could mean for example that every tenth frame is selected as a key frame. In one aspect, key frames may be selected as data is acquired based on any number of suitable criteria. In another aspect, key frame pairs may be determined post hoc by examining all possible candidate key frames. All possible key frame pairs may be examined and candidates may be removed, for example, where there is insufficient overlap to form a stitch. Still more generally, any technique suitable for selecting a subset of frames in a data set may be usefully employed to select key frames for processing in order to reduce computational complexity.
[0055] Once key frames have been selected, additional processing may be performed. For example, full image data (e.g., full resolution center and side channel images) may be stored for each key frame, along with image signature data, point cloud centroid calculations, and any other measured or calculated data to support use of the key frames in a three-dimensional reconstruction process as described herein.
[0056] As shown in step 714, candidate stitches may be identified. In general, a stitch is a relationship between two separate three-dimensional measurements from two different camera poses. Once a stitch is established, a rotation and a translation may be determined for the path of a camera between the two poses. In a complementary fashion, the three-dimensional measurements from the poses may be combined into a portion of a three-dimensional model. Candidate stitches may be analyzed around each key frame, such as from the key frame to some or all of the frames of data between the key frame and neighboring key frames. In another aspect, a candidate stitch may be made to every other key frame, or in order to reduce computational complexity, every key frame within a spatial or sequential neighborhood around a key frame. Stitches may be based on the originally imaged frames. It may also be useful to deform or warp two-dimensional images during registration and other steps in a stitching process in order to improve accuracy and/or speed. Stitches may also or instead be based on other observed epipolar relationships in source data. [0057] As shown in step 716, stitches may be selected for the complete camera path from the universe of candidate stitches. The selection of stitches may be made based upon, e.g., the lowest calculated error in resulting portions of the three-dimensional model. In general, each key frame may be stitched to one or more other key frames and each non-key frame may be stitched to at least one sequentially neighboring key frame.
[0058] As shown in step 718, a graph analysis may be performed using the key frames and the associated stitching to calculate a global path for the camera used to obtain a three-dimensional model. The graph analysis may consider each key frame as a node or vertex and each stitch as an edge between a pair of nodes. A key frame is selected as a starting point. A breadth- or depth- first search may be performed through the graph to identify stitches which may connect the current key frame to another key frame. Each key frame may be marked as the graph is processed. A check may be performed to see if all key frames have been reached within the graph. If all key frames have not been reached through traversing stitches in the graph analysis, the largest subgraph is identified. This sub-graph may be examined to see if the entire three- dimensional image may be modeled.
[0059] It may be that certain sub-graphs are not required to complete the three- dimensional imaging. If the camera lingered over a particular region of a surface of an object, or if the camera looped on a region multiple times, the associated sub-graph(s) may not be needed. If a separate sub-graph is identified, which is needed to complete the three-dimensional imaging, an optional branch back to step 712 may be performed. For example, a set of key frames may have been selected which did not have sufficient stitching from one key frame to the next key frame. By choosing a different set of key frames, sufficient stitching may be obtained in order to obtain a complete graph of all needed aspects of the three-dimensional imaging. A key frame which is too sparse, meaning it has insufficient stitches to aid in building a graph, may indicate that a different set of key frames should be selected. Based on the graph analysis, a global path may be selected, and the graph may then be analyzed to optimize the path calculation.
[0060] As shown in step 720, a numerical optimization may be performed to reduce errors in the calculated camera path based upon available data for the complete camera path such as, for example, cross links that interrelate temporally distant measurements. In general, the objective of numerical optimization is to minimize a calculated error based upon an error function for the camera path and/or reconstructed three-dimensional model. A useful formulation of the error minimization problem for a global camera path is presented below.
[0061] In general, a set of related measurements in a path may be captured from three poses, A, B, and C, each related to one another and to a world coordinate system having an origin, O, by motion parameters of rotation and a translation. The relationship between a point, X, expressed in the world coordinate system as Xo and the same point expressed in the A coordinate system, XA may be expressed as:
XA = R0AX0 + T0A [Eq. 1]
R is the rotation taking points from the world to the A coordinate system. TOA is the translation of the world coordinate system to the A coordinate system. It should be understood that symbols X and T may represent a vector, rather than a scalar, e.g. where X includes x, y, and z coordinate values. Further, it should be understood that symbol R may represent a matrix. The following equations may similarly represent the transformation between the world and the B and C coordinate systems respectively: XB = ROBXO + T0B [Eq. 2]
Xc = R0CX0 + Toc [Eq. 3]
[0062] By rearranging, equations 1 and 2 may be represented as follows: X0 = R0 ~ A (XA - T0A )= R0 ~ B (XB - T0B ) [Eq. 4]
[0063] The representation of a point in one camera's coordinate system may be related to the same point in another coordinate system. For example, as in equations 1-3, coordinates of a point, X, may be transformed from the A coordinate system to the B coordinate system as follows:
XB = RABXA + TAB [Eq. 5]
[0064] The rotation RAB rotates points from the A to the B coordinate system and TAB translates the origin of the A coordinate system in the B coordinate system.
[0065] In optimization, the pose of every camera may be optimized based on measured transformations between poses. That is, a number of camera-to-world rotations and camera-to-world translations, Ron and Ton may be performed. In general, one of these may be defined as the identity rotation and zero translation, with the remaining values being optimized as described below.
[0066] The rotations and translations may be measured for many pairs of cameras. For the zth such measured frame-to-frame motion, let one of the cameras of the pair be camera A and the other be camera B. This may also be considered the zth stitch. Let RAB be the measured rotation taking points in the A system to the B system and TAB be the coordinates of the A position expressed in the B system, as in equation 5.
[0067] The rotations and translations for all cameras, Ron and Ton may be optimized. Rc' 0A and Rc' 0B may be defined to be the candidate rotations; Tc' 0A and
Tc l 0B may be defined to be the candidate translations corresponding to the A and B camera of the zth stitch. Further, Rc' AB = Rc' 0B (Rc' 0A )~l may be defined as the candidate rotation from A to B, and Tc l AB = Tc l 0B - Rc' MT^ 0A , the candidate translation for the transformation from A to B.
[0068] Note that with sufficient stitches, the motion constraints may form an overdetermined system of motion constraint equations. Using these equations as a starting point, numerical optimization may be performed on the rotational and translational components of each camera based on the measured stitches.
[0069] In a decoupled optimization, the rotational and translational components may be independently optimized. Given a candidate set of camera rotations, Rc' the corresponding candidate camera-to-camera rotations, Rc' AB , may be computed that correspond to each of the measured camera-to-camera rotations, RAB . Thus the corresponding residual rotations are given by Rr ι esιdual AB = Rc' AB (RA'B )~1 • A scalar- valued rotational cost function, er, may be computed that depends on the candidate camera rotations
#stιtches er (RC,O» ) = Σ K K > where rr = %SOO) Rr1eS,dual,AB > [EΦ 6]
[0070] In equation 6, logso(3)(R) returns the axis-angle vector, v, that corresponds to the rotation R. In other words, logso(3)(R) returns the vector, v, that has a cross-product matrix, [v]x , that is the matrix logarithm of R. [0071] Next, a similar scalar-valued cost function may be computed for translation that depends on the candidate rotations and translations.
et (Rc,on , TCfin ) = #Sψ rfr] , where r] = T^ - TAB [Eq. 7]
[0072] Equation 6 may be minimized as a nonlinear optimization; equation 7 may be minimized as a linear optimization.
[0073] In one conventional, decoupled approach to solving these simultaneous systems of equations, the rotational error function may be converted into a quaternion expression in order to translate the numerical problem into a linear system of equations for solution. While this approach may increase computational efficiency, it offers an incomplete optimization solution.
[0074] The decoupled approach described above does not provide a truly optimal one, in a maximum-likelihood sense, as it cannot use information from the translation portion of the stitches in determining rotation. In order to achieve a coupled optimization a weighting may be used to balance the contributions of rotational and translational components to a combined cost function:
ec(RC,On Jc,On ) = [Eq. 8]
Figure imgf000026_0001
Multiple approaches may be used to optimize this cost function, but in one embodiment the weights may be expressed as matrices. Different stitches may receive different weightings based upon a number of factors including the number of points in the stitch (e.g., the shared content), the quality of a particular three-dimensional measurement, and/or any other factors impacting the known reliability of a stitch. In one approach, the weight matrices may also account for anisotropic error in the individual points collected, such as due to acquisition of depth information from disparity measurements, which results in measurement precision that varies with distance from the camera.
[0075] In some cases, equation 8 may be reformulated so that the rotation and translation weights are decoupled for each stitch (i.e., Wc' is a block diagonal). In particular, this may occur in the case where the motion stitches are recovered from three- dimensional point correspondences with isotropic point error. In that case, for a given stitch i, between camera A and camera B, the optimal solution may bring the point cloud as seen from camera A into correspondence with that seen from camera B. If XA' and XB' are the positions of the center of the point cloud in the A and B systems respectively, then rt' may be replaced in equation 8 with the residual displacement between the point-cloud centers based on the candidate camera pose as follows:
<ctr = XB' - (Rc' ,ABXA' + T^AB) [Eq. 9]
Equation 8 may then be reformulated as:
ec(Rc,onJc,on) = **T (//CC + rfwrf' ) [Eq. 10]
This coupled optimization problem may still be considered as being non-linear. It should be understood that other optimizations are also possible that would fall within the scope of this disclosure.
[0076] In general, by minimizing equation 8, both rotational errors and translational errors may be minimized simultaneously. The weight matrices can be chosen, for example, according to "First Order Error Propagation of the Procrustes Method for 3D Attitude Estimation" by Leo Dorst, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 27, No. 2, Feb. 2005, pp. 221-9 which is incorporated in its entirety by reference. Once a more consistent set of motion parameters has been generated the three-dimensional model may be updated.
[0077] In one aspect, the residual error may be employed as a calibration metric. When total error or some portion of error has been minimized, the residual error may be evaluated. If a minimized error falls beyond a certain threshold then calibration for the camera and associated hardware may be recommended, based upon an inference that the inability to produce better quality results is due to a miscalibration or other malfunction of the camera system. The threshold value may be empirically determined based on the specific camera hardware equipment or it may be learned experientially over time for a given system. When a system is new or has been freshly aligned, expected minimized error values may be obtained. When minimized error values deviate from these expected values, a calibration state evaluation flag may be set, or other alert or message generated, indicating that the tool should be calibrated. [0078] As shown in step 722, upsampling may be performed to augment a three- dimensional model with data from non-key frames. For example, non-key frames may be registered to nearby key frames to create small, local reconstruction patches including the full image detail available from non-key frames. In this manner, path optimization may be performed on a key- frame-based data set, thus reducing the data requiring processing, while retaining additional data points from non-key frames for use in the final three- dimensional model.
[0079] It will be appreciated that any of the above system and/or methods may be realized in hardware, software, or any combination of these suitable for the data acquisition and modeling technologies described herein. This includes realization in one or more microprocessors, microcontrollers, embedded microcontrollers, programmable digital signal processors or other programmable devices, along with internal and/or external memory. The may also, or instead, include one or more application specific integrated circuits, programmable gate arrays, programmable array logic components, or any other device or devices that may be configured to process electronic signals. It will further be appreciated that a realization may include computer executable code created using a structured programming language such as C, an object oriented programming language such as C++, or any other high-level or low-level programming language (including assembly languages, hardware description languages, and database programming languages and technologies) that may be stored, compiled or interpreted to run on one of the above devices, as well as heterogeneous combinations of processors, processor architectures, or combinations of different hardware and software. Thus in one aspect there is disclosed herein a computer program product comprising computer executable code that, when executing on one or more computing devices, performs any and/or all of the steps described above. At the same time, processing may be distributed across devices such as a camera and/or computer and/or fabrication facility and/or dental laboratory and/or server in a number of ways or all of the functionality may be integrated into a dedicated, standalone device. All such permutations and combinations are intended to fall within the scope of the present disclosure.

Claims

CLAIMSWhat is claimed is:
1. A method of three-dimensional reconstruction comprising: acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; selecting a subset of the plurality of frames of image data to provide a plurality of key frames, each one of the plurality of key frames related to at least one other one of the plurality of key frames by a portion of the camera path including a rotation and a translation determined based upon one or more common points in the three-dimensional reconstruction of the surface of the object in each of the respective key frames, wherein the remaining ones of the plurality of frames of image data are non-key frames; providing a three-dimensional model of the object; determining a second rotation and a second translation from one of the key frames to at least one of the non-key frames sequentially positioned between the one of the key frames and a sequentially adjacent one of the key frames; obtaining three-dimensional reconstruction information of the surface of the object from the camera position of the at least one of the non-key frames to provide upsampled three-dimensional data; and adding the upsampled three-dimensional data to the three-dimensional model based upon the second rotation and the second translation.
2. The method according to claim 1 further comprising estimating a camera motion between two adjacent key frames wherein the estimating is based on the rotation and the translation.
3. The method according to claim 2 further comprising optimizing the estimation of the camera motion between the two adjacent key frames by creating consistency among motion parameters using an overdetermined system of motion constraint equations, wherein the motion parameters are comprised of information on the rotation and the translation.
4. The method according to claim 3 further comprising optimizing the camera motion between two adjacent non-key frames by creating consistency among motion parameters using an overdetermined system of motion constraint equations.
5. The method according to claim 4 further comprising updating the three- dimensional reconstruction based on the created consistency among the motion parameters.
6. The method according to claim 1 wherein the data for a three-dimensional reconstruction of the surface of the object was obtained from at least one other channel image to provide disparity data.
7. The method according to claim 1 wherein providing a three-dimensional model further comprises generating a three-dimensional model of the object using the camera path and the three-dimensional reconstruction for each of the key frames.
8. The method according to claim 1 further comprising obtaining three-dimensional reconstruction information of the surface of the object from the camera position for all of the non-key frames between two adjacent key frames.
9. The method according to claim 1 wherein the selecting of the subset of the plurality of frames is based on a quality metric of the three-dimensional reconstruction.
10. The method according to claim 1 wherein the selecting of the subset of the plurality of frames is determined using graph analysis to ensure that all of the key frames are utilized in the three-dimensional reconstruction.
11. A computer program product comprising computer executable code embodied in a computer readable medium that, when executing on one or more computing devices, performs the steps of: acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; selecting a subset of the plurality of frames of image data to provide a plurality of key frames, each one of the plurality of key frames related to at least one other one of the plurality of key frames by a portion of the camera path including a rotation and a translation determined based upon one or more common points in the three-dimensional reconstruction of the surface of the object in each of the respective key frames, wherein the remaining ones of the plurality of frames of image data are non-key frames; providing a three-dimensional model of the object; determining a second rotation and a second translation from one of the key frames to at least one of the non-key frames sequentially positioned between the one of the key frames and a sequentially adjacent one of the key frames; obtaining three-dimensional reconstruction information of the surface of the object from the camera position of the at least one of the non-key frames to provide upsampled three-dimensional data; and adding the upsampled three-dimensional data to the three-dimensional model based upon the second rotation and the second translation.
12. The computer program product of claim 11 further comprising computer executable code that performs the step of estimating a camera motion between two adjacent key frames wherein the estimating is based on the rotation and the translation.
13. The computer program product of claim 12 further comprising computer executable code that performs the step of optimizing the estimation of the camera motion between the two adjacent key frames by creating consistency among motion parameters using an overdetermined system of motion constraint equations, wherein the motion parameters are comprised of information on the rotation and the translation.
14. The computer program product of claim 13 further comprising computer executable code that performs the step of optimizing the camera motion between two adjacent non-key frames by creating consistency among motion parameters using an overdetermined system of motion constraint equations.
15. The computer program product of claim 14 further comprising computer executable code that performs the step of updating the three-dimensional reconstruction based on the created consistency among the motion parameters.
16. The computer program product of claim 11 wherein the data for a three- dimensional reconstruction of the surface of the object was obtained from at least one other channel image to provide disparity data.
17. The computer program product of claim 11 wherein providing a three- dimensional model further comprises generating a three-dimensional model of the object using the camera path and the three-dimensional reconstruction for each of the key frames.
18. The computer program product of claim 11 further comprising computer executable code that performs the step of obtaining three-dimensional reconstruction information of the surface of the object from the camera position for all of the non-key frames between two adjacent key frames.
19. The computer program product of claim 11 wherein the selecting of the subset of the plurality of frames is based on a quality metric of the three-dimensional reconstruction.
20. The computer program product of claim 11 wherein the selecting of the subset of the plurality of frames is determined using graph analysis to ensure that all of the key frames are utilized in the three-dimensional reconstruction.
21. A method for interactively reducing accumulated error in a global path, the method comprising: acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; generating a three-dimensional model of the object using the camera path and the data for the three-dimensional reconstruction; identifying two of the plurality of frames of image data that represent a candidate for an accumulated error in the camera path relative to one another; and displaying the three-dimensional model along with a graphical annotation that illustrates a recommended scan path to reduce the accumulated error.
22. The method according to claim 21 wherein the data for a three-dimensional reconstruction of the surface of the object was obtained from at least one other channel image to provide disparity data.
23. The method according to claim 21 further comprising acquiring one or more frames of image data along the recommended scan path to reduce the accumulated error.
24. The method according to claim 21 wherein identifying two of the plurality of frames further comprises identifying frames of image data that are separated by a substantially greater distance along the camera path than along the surface of the object.
25. A computer program product comprising computer executable code embodied in a computer readable medium that, when executing on one or more computing devices, performs the steps of: acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; generating a three-dimensional model of the object using the camera path and the data for the three-dimensional reconstruction; identifying two of the plurality of frames of image data that represent a candidate for an accumulated error in the camera path relative to one another; and displaying the three-dimensional model along with a graphical annotation that illustrates a recommended scan path to reduce the accumulated error.
26. The computer program product of claim 25 wherein the data for a three- dimensional reconstruction of the surface of the object was obtained from at least one other channel image to provide disparity data.
27. The computer program product of claim 25 further comprising computer executable code that performs the step of acquiring one or more frames of image data along the recommended scan path to reduce the accumulated error.
28. The computer program product of claim 25 wherein identifying two of the plurality of frames further comprises identifying frames of image data that are separated by a substantially greater distance along the camera path than along the surface of the object.
29. A system comprising a camera, a monitor, a processor, and a memory, the memory storing a computer program executable by the processor to perform the steps of: acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; generating a three-dimensional model of the object using the camera path and the data for the three-dimensional reconstruction; identifying two of the plurality of frames of image data that represent a candidate for an accumulated error in the camera path relative to one another; and displaying the three-dimensional model along with a graphical annotation that illustrates a recommended scan path to reduce the accumulated error.
30. The system of claim 29 wherein the data for a three-dimensional reconstruction of the surface of the object was obtained from at least one other channel image to provide disparity data.
31. The system of claim 29, the memory storing computer program product of claim 29 further comprising computer executable code that performs the step of acquiring one or more frames of image data along the recommended scan path to reduce the accumulated error.
32. The system of claim 29 wherein identifying two of the plurality of frames further comprises identifying frames of image data that are separated by a substantially greater distance along the camera path than along the surface of the object.
33. A method for global path optimization, the method comprising: acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; and minimizing an error function for a plurality of camera positions along the camera path, the error function including a system of equations for translational components of an error and for rotational components of the error, wherein the error function couples the translational components and the rotational components using a weighting matrix, thereby providing an optimized camera path.
34. The method according to claim 33 wherein the data for a three-dimensional reconstruction of the surface of the object was obtained from at least one other channel image to provide disparity data.
35. The method according to claim 33 wherein the system of equations is a non- linear system of equations.
36. The method according to claim 35 wherein the translational component of the error forms of a system of linear equations.
37. The method according to claim 35 wherein the rotational component of the error forms a system of non- linear equations.
38. The method according to claim 33 further comprising generating a three- dimensional model based upon the camera path and the data for the three-dimensional reconstruction, and refining the three-dimensional model based upon the optimized camera path.
39. The method according to claim 33 further comprising selecting a subset of the plurality of frames of image data to provide a plurality of key frames, each one of the plurality of key frames related to at least one other one of the plurality of key frames by a portion of the camera path including a rotation and a translation determined based upon one or more common points in the three-dimensional reconstruction of the surface of the object in each of the respective key frames, wherein the remaining ones of the plurality of frames of image data are non-key frames.
40. The method according to claim 33 wherein the weighting matrix is selected to locally decouple the error function around a centroid of common surface data for two or more three-dimensional reconstructions.
41. The method according to claim 33 wherein minimizing an error function further comprises evaluating a calibration state based on the resulting error function minimization.
42. A computer program product comprising computer executable code embodied in a computer readable medium that, when executing on one or more computing devices, performs the steps of: acquiring a plurality of frames of image data of a surface of an object, each one of the plurality of frames of image data captured from a camera position along a camera path and each one of the plurality of frames of image data including a conventional image of the object from the camera position and data for a three-dimensional reconstruction of the surface of the object as viewed from the camera position; and minimizing an error function for a plurality of camera positions along the camera path, the error function including a system of equations for translational components of an error and for rotational components of the error, wherein the error function couples the translational components and the rotational components using a weighting matrix, thereby providing an optimized camera path.
43. The computer program product of claim 42 wherein the data for a three- dimensional reconstruction of the surface of the object was obtained from at least one other channel image to provide disparity data.
44. The computer program product of claim 42 wherein the system of equations is a non-linear system of equations.
45. The computer program product of claim 44 wherein the translational component of the error forms of a system of linear equations.
46. The computer program product of claim 44 wherein the rotational component of the error forms a system of non-linear equations.
47. The computer program product of claim 42 further comprising computer executable code that performs the step of generating a three-dimensional model based upon the camera path and the data for the three-dimensional reconstruction, and refining the three-dimensional model based upon the optimized camera path.
48. The computer program product of claim 42 further comprising computer executable code that performs the step of selecting a subset of the plurality of frames of image data to provide a plurality of key frames, each one of the plurality of key frames related to at least one other one of the plurality of key frames by a portion of the camera path including a rotation and a translation determined based upon one or more common points in the three-dimensional reconstruction of the surface of the object in each of the respective key frames, wherein the remaining ones of the plurality of frames of image data are non-key frames.
49. The computer program product of claim 42 wherein the weighting matrix is selected to locally decouple the error function around a centroid of common surface data for two or more three-dimensional reconstructions.
50. The computer program product of claim 42 wherein minimizing an error function further comprises evaluating a calibration state based on the resulting error function minimization.
PCT/US2009/030068 2008-01-04 2009-01-04 Global camera path optimization WO2009089129A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US12/811,239 US8803958B2 (en) 2008-01-04 2009-01-04 Global camera path optimization
DE112009000101T DE112009000101T5 (en) 2008-01-04 2009-01-04 Global camera path optimization

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US1915908P 2008-01-04 2008-01-04
US61/019,159 2008-01-04

Publications (1)

Publication Number Publication Date
WO2009089129A1 true WO2009089129A1 (en) 2009-07-16

Family

ID=40853423

Family Applications (5)

Application Number Title Priority Date Filing Date
PCT/US2009/030067 WO2009089128A1 (en) 2008-01-04 2009-01-04 Hierarchical processing using image deformation
PCT/US2009/030066 WO2009089127A1 (en) 2008-01-04 2009-01-04 Image signatures for use in motion-based three-dimensional reconstruction
PCT/US2009/030065 WO2009089126A1 (en) 2008-01-04 2009-01-04 Three-dimensional model refinement
PCT/US2009/030068 WO2009089129A1 (en) 2008-01-04 2009-01-04 Global camera path optimization
PCT/US2009/030064 WO2009089125A2 (en) 2008-01-04 2009-01-04 Navigating among images of an object in 3d space

Family Applications Before (3)

Application Number Title Priority Date Filing Date
PCT/US2009/030067 WO2009089128A1 (en) 2008-01-04 2009-01-04 Hierarchical processing using image deformation
PCT/US2009/030066 WO2009089127A1 (en) 2008-01-04 2009-01-04 Image signatures for use in motion-based three-dimensional reconstruction
PCT/US2009/030065 WO2009089126A1 (en) 2008-01-04 2009-01-04 Three-dimensional model refinement

Family Applications After (1)

Application Number Title Priority Date Filing Date
PCT/US2009/030064 WO2009089125A2 (en) 2008-01-04 2009-01-04 Navigating among images of an object in 3d space

Country Status (3)

Country Link
US (7) US8803958B2 (en)
DE (5) DE112009000099T5 (en)
WO (5) WO2009089128A1 (en)

Cited By (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150257853A1 (en) 2009-02-02 2015-09-17 Viax Dental Technologies, LLC Dentist tool
WO2016120767A1 (en) * 2015-01-27 2016-08-04 Align Technology, Inc. Method and system for oral-cavity-imaging-and-modeling equipment
US10123706B2 (en) 2016-07-27 2018-11-13 Align Technology, Inc. Intraoral scanner with dental diagnostics capabilities
US10130445B2 (en) 2014-09-19 2018-11-20 Align Technology, Inc. Arch expanding appliance
US10144100B2 (en) 2009-02-02 2018-12-04 Viax Dental Technologies, LLC Method of preparation for restoring tooth structure
US10248883B2 (en) 2015-08-20 2019-04-02 Align Technology, Inc. Photograph-based assessment of dental treatments and procedures
US10327872B2 (en) 2014-08-15 2019-06-25 Align Technology, Inc. Field curvature model for confocal imaging apparatus with curved focal surface
US10383705B2 (en) 2016-06-17 2019-08-20 Align Technology, Inc. Orthodontic appliance performance monitor
US10390913B2 (en) 2018-01-26 2019-08-27 Align Technology, Inc. Diagnostic intraoral scanning
US10421152B2 (en) 2011-09-21 2019-09-24 Align Technology, Inc. Laser cutting
US10426572B2 (en) 2011-05-26 2019-10-01 Viax Dental Technologies Llc Dental tool and guidance devices
US10449016B2 (en) 2014-09-19 2019-10-22 Align Technology, Inc. Arch adjustment appliance
US10456043B2 (en) 2017-01-12 2019-10-29 Align Technology, Inc. Compact confocal dental scanning apparatus
US10470847B2 (en) 2016-06-17 2019-11-12 Align Technology, Inc. Intraoral appliances with sensing
US10507087B2 (en) 2016-07-27 2019-12-17 Align Technology, Inc. Methods and apparatuses for forming a three-dimensional volumetric model of a subject's teeth
US10517482B2 (en) 2017-07-27 2019-12-31 Align Technology, Inc. Optical coherence tomography for orthodontic aligners
US10524881B2 (en) 2010-04-30 2020-01-07 Align Technology, Inc. Patterned dental positioning appliance
US10537405B2 (en) 2014-11-13 2020-01-21 Align Technology, Inc. Dental appliance with cavity for an unerupted or erupting tooth
US10543064B2 (en) 2008-05-23 2020-01-28 Align Technology, Inc. Dental implant positioning
US10548700B2 (en) 2016-12-16 2020-02-04 Align Technology, Inc. Dental appliance etch template
US10595966B2 (en) 2016-11-04 2020-03-24 Align Technology, Inc. Methods and apparatuses for dental images
US10610332B2 (en) 2012-05-22 2020-04-07 Align Technology, Inc. Adjustment of tooth position in a virtual dental model
US10613515B2 (en) 2017-03-31 2020-04-07 Align Technology, Inc. Orthodontic appliances including at least partially un-erupted teeth and method of forming them
US10639134B2 (en) 2017-06-26 2020-05-05 Align Technology, Inc. Biosensor performance indicator for intraoral appliances
US10758321B2 (en) 2008-05-23 2020-09-01 Align Technology, Inc. Smile designer
US10772506B2 (en) 2014-07-07 2020-09-15 Align Technology, Inc. Apparatus for dental confocal imaging
US10779718B2 (en) 2017-02-13 2020-09-22 Align Technology, Inc. Cheek retractor and mobile device holder
US10813720B2 (en) 2017-10-05 2020-10-27 Align Technology, Inc. Interproximal reduction templates
US10842601B2 (en) 2008-06-12 2020-11-24 Align Technology, Inc. Dental appliance
US10885521B2 (en) 2017-07-17 2021-01-05 Align Technology, Inc. Method and apparatuses for interactive ordering of dental aligners
US10893918B2 (en) 2012-03-01 2021-01-19 Align Technology, Inc. Determining a dental treatment difficulty
US10919209B2 (en) 2009-08-13 2021-02-16 Align Technology, Inc. Method of forming a dental appliance
US10980613B2 (en) 2017-12-29 2021-04-20 Align Technology, Inc. Augmented reality enhancements for dental practitioners
US10993783B2 (en) 2016-12-02 2021-05-04 Align Technology, Inc. Methods and apparatuses for customizing a rapid palatal expander
US11007035B2 (en) 2017-03-16 2021-05-18 Viax Dental Technologies Llc System for preparing teeth for the placement of veneers
US11026831B2 (en) 2016-12-02 2021-06-08 Align Technology, Inc. Dental appliance features for speech enhancement
US11026768B2 (en) 1998-10-08 2021-06-08 Align Technology, Inc. Dental appliance reinforcement
US11045283B2 (en) 2017-06-09 2021-06-29 Align Technology, Inc. Palatal expander with skeletal anchorage devices
US11083545B2 (en) 2009-03-19 2021-08-10 Align Technology, Inc. Dental wire attachment
US11096763B2 (en) 2017-11-01 2021-08-24 Align Technology, Inc. Automatic treatment planning
US11103330B2 (en) 2015-12-09 2021-08-31 Align Technology, Inc. Dental attachment placement structure
US11116605B2 (en) 2017-08-15 2021-09-14 Align Technology, Inc. Buccal corridor assessment and computation
US11123156B2 (en) 2017-08-17 2021-09-21 Align Technology, Inc. Dental appliance compliance monitoring
US11213368B2 (en) 2008-03-25 2022-01-04 Align Technology, Inc. Reconstruction of non-visible part of tooth
US11219506B2 (en) 2017-11-30 2022-01-11 Align Technology, Inc. Sensors for monitoring oral appliances
US11273011B2 (en) 2016-12-02 2022-03-15 Align Technology, Inc. Palatal expanders and methods of expanding a palate
US11376101B2 (en) 2016-12-02 2022-07-05 Align Technology, Inc. Force control, stop mechanism, regulating structure of removable arch adjustment appliance
US11419702B2 (en) 2017-07-21 2022-08-23 Align Technology, Inc. Palatal contour anchorage
US11426259B2 (en) 2012-02-02 2022-08-30 Align Technology, Inc. Identifying forces on a tooth
US11432908B2 (en) 2017-12-15 2022-09-06 Align Technology, Inc. Closed loop adaptive orthodontic treatment methods and apparatuses
US11471252B2 (en) 2008-10-08 2022-10-18 Align Technology, Inc. Dental positioning appliance having mesh portion
US11534974B2 (en) 2017-11-17 2022-12-27 Align Technology, Inc. Customized fabrication of orthodontic retainers based on patient anatomy
US11534268B2 (en) 2017-10-27 2022-12-27 Align Technology, Inc. Alternative bite adjustment structures
US11554000B2 (en) 2015-11-12 2023-01-17 Align Technology, Inc. Dental attachment formation structure
US11576752B2 (en) 2017-10-31 2023-02-14 Align Technology, Inc. Dental appliance having selective occlusal loading and controlled intercuspation
US11596502B2 (en) 2015-12-09 2023-03-07 Align Technology, Inc. Dental attachment placement structure
US11612454B2 (en) 2010-04-30 2023-03-28 Align Technology, Inc. Individualized orthodontic treatment index
US11633268B2 (en) 2017-07-27 2023-04-25 Align Technology, Inc. Tooth shading, transparency and glazing
US11931222B2 (en) 2015-11-12 2024-03-19 Align Technology, Inc. Dental attachment formation structures
US11937991B2 (en) 2018-03-27 2024-03-26 Align Technology, Inc. Dental attachment placement structure

Families Citing this family (186)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006092800A2 (en) * 2005-03-03 2006-09-08 Cadent Ltd. System and method for scanning an intraoral cavity
US11403336B2 (en) 2005-10-26 2022-08-02 Cortica Ltd. System and method for removing contextually identical multimedia content elements
US11003706B2 (en) 2005-10-26 2021-05-11 Cortica Ltd System and methods for determining access permissions on personalized clusters of multimedia content elements
US9372940B2 (en) 2005-10-26 2016-06-21 Cortica, Ltd. Apparatus and method for determining user attention using a deep-content-classification (DCC) system
US10387914B2 (en) 2005-10-26 2019-08-20 Cortica, Ltd. Method for identification of multimedia content elements and adding advertising content respective thereof
US10742340B2 (en) 2005-10-26 2020-08-11 Cortica Ltd. System and method for identifying the context of multimedia content elements displayed in a web-page and providing contextual filters respective thereto
US11032017B2 (en) 2005-10-26 2021-06-08 Cortica, Ltd. System and method for identifying the context of multimedia content elements
US10691642B2 (en) 2005-10-26 2020-06-23 Cortica Ltd System and method for enriching a concept database with homogenous concepts
US10848590B2 (en) 2005-10-26 2020-11-24 Cortica Ltd System and method for determining a contextual insight and providing recommendations based thereon
US10372746B2 (en) 2005-10-26 2019-08-06 Cortica, Ltd. System and method for searching applications using multimedia content elements
US9384196B2 (en) 2005-10-26 2016-07-05 Cortica, Ltd. Signature generation for multimedia deep-content-classification by a large-scale matching system and method thereof
US11604847B2 (en) 2005-10-26 2023-03-14 Cortica Ltd. System and method for overlaying content on a multimedia content element based on user interest
US11216498B2 (en) * 2005-10-26 2022-01-04 Cortica, Ltd. System and method for generating signatures to three-dimensional multimedia data elements
US10380267B2 (en) 2005-10-26 2019-08-13 Cortica, Ltd. System and method for tagging multimedia content elements
US10607355B2 (en) 2005-10-26 2020-03-31 Cortica, Ltd. Method and system for determining the dimensions of an object shown in a multimedia content item
US11019161B2 (en) 2005-10-26 2021-05-25 Cortica, Ltd. System and method for profiling users interest based on multimedia content analysis
US10193990B2 (en) 2005-10-26 2019-01-29 Cortica Ltd. System and method for creating user profiles based on multimedia content
US10585934B2 (en) 2005-10-26 2020-03-10 Cortica Ltd. Method and system for populating a concept database with respect to user identifiers
US10621988B2 (en) 2005-10-26 2020-04-14 Cortica Ltd System and method for speech to text translation using cores of a natural liquid architecture system
US9646005B2 (en) 2005-10-26 2017-05-09 Cortica, Ltd. System and method for creating a database of multimedia content elements assigned to users
US10180942B2 (en) 2005-10-26 2019-01-15 Cortica Ltd. System and method for generation of concept structures based on sub-concepts
US10614626B2 (en) 2005-10-26 2020-04-07 Cortica Ltd. System and method for providing augmented reality challenges
US8818916B2 (en) 2005-10-26 2014-08-26 Cortica, Ltd. System and method for linking multimedia data elements to web pages
US10776585B2 (en) 2005-10-26 2020-09-15 Cortica, Ltd. System and method for recognizing characters in multimedia content
US20160321253A1 (en) 2005-10-26 2016-11-03 Cortica, Ltd. System and method for providing recommendations based on user profiles
US10380623B2 (en) 2005-10-26 2019-08-13 Cortica, Ltd. System and method for generating an advertisement effectiveness performance score
US8326775B2 (en) 2005-10-26 2012-12-04 Cortica Ltd. Signature generation for multimedia deep-content-classification by a large-scale matching system and method thereof
US10733326B2 (en) 2006-10-26 2020-08-04 Cortica Ltd. System and method for identification of inappropriate multimedia content
DE112009000099T5 (en) 2008-01-04 2010-11-11 3M Innovative Properties Co., St. Paul Image signatures for use in a motion-based three-dimensional reconstruction
US9094675B2 (en) * 2008-02-29 2015-07-28 Disney Enterprises Inc. Processing image data from multiple cameras for motion pictures
US11792538B2 (en) 2008-05-20 2023-10-17 Adeia Imaging Llc Capturing and processing of images including occlusions focused on an image sensor by a lens stack array
US8866920B2 (en) 2008-05-20 2014-10-21 Pelican Imaging Corporation Capturing and processing of images using monolithic camera array with heterogeneous imagers
JP5654583B2 (en) 2009-06-17 2015-01-14 3シェイプ アー/エス Focus control device
WO2011034781A2 (en) 2009-09-15 2011-03-24 3M Innovative Properties Company Method of making dental implant model and articles
WO2011034780A1 (en) 2009-09-15 2011-03-24 3M Innovative Properties Company Dental implant abutments and methods of use
AT508563B1 (en) * 2009-10-07 2011-02-15 Ait Austrian Inst Technology METHOD FOR RECORDING THREE-DIMENSIONAL IMAGES
DE102009048834A1 (en) * 2009-10-09 2011-04-14 Volkswagen Ag Method and display device for displaying information
US8566020B2 (en) * 2009-12-01 2013-10-22 Nokia Corporation Method and apparatus for transforming three-dimensional map objects to present navigation information
US8624902B2 (en) 2010-02-04 2014-01-07 Microsoft Corporation Transitioning between top-down maps and local navigation of reconstructed 3-D scenes
US8773424B2 (en) 2010-02-04 2014-07-08 Microsoft Corporation User interfaces for interacting with top-down maps of reconstructed 3-D scences
US8295589B2 (en) 2010-05-20 2012-10-23 Microsoft Corporation Spatially registering user photographs
JP2012053268A (en) * 2010-09-01 2012-03-15 Canon Inc Lenticular lens, image forming apparatus and image forming method
WO2012061549A2 (en) * 2010-11-03 2012-05-10 3Dmedia Corporation Methods, systems, and computer program products for creating three-dimensional video sequences
KR101791590B1 (en) * 2010-11-05 2017-10-30 삼성전자주식회사 Object pose recognition apparatus and method using the same
WO2012076013A1 (en) 2010-12-06 2012-06-14 3Shape A/S System with 3d user interface integration
ES2788853T3 (en) 2010-12-06 2020-10-23 3Shape As System with 3D user interface integration
US8878950B2 (en) 2010-12-14 2014-11-04 Pelican Imaging Corporation Systems and methods for synthesizing high resolution images using super-resolution processes
EP2654606B1 (en) * 2010-12-21 2021-05-26 3Shape A/S Motion blur compensation
US20120162394A1 (en) * 2010-12-23 2012-06-28 Tektronix, Inc. Displays for easy visualizing of 3d disparity data
EP2678830B1 (en) * 2011-02-22 2020-04-08 Midmark Corporation Hybrid stitching for 3d reconstruction
US8774455B2 (en) 2011-03-02 2014-07-08 Raf Technology, Inc. Document fingerprinting
US9443298B2 (en) 2012-03-02 2016-09-13 Authentect, Inc. Digital fingerprinting object authentication and anti-counterfeiting system
JP5144784B2 (en) * 2011-04-11 2013-02-13 ファナック株式会社 Tool trajectory display device for machine tools
US9433795B2 (en) 2011-04-13 2016-09-06 Boston Scientific Neuromodulation Corporation User interface with view finder for localizing anatomical region
US8963998B2 (en) * 2011-04-15 2015-02-24 Tektronix, Inc. Full reference system for predicting subjective quality of three-dimensional video
WO2012168322A2 (en) 2011-06-06 2012-12-13 3Shape A/S Dual-resolution 3d scanner
US8648919B2 (en) * 2011-06-06 2014-02-11 Apple Inc. Methods and systems for image stabilization
US8823813B2 (en) 2011-06-06 2014-09-02 Apple Inc. Correcting rolling shutter using image stabilization
US20120329008A1 (en) * 2011-06-22 2012-12-27 Trident Labs, Inc. d/b/a Trident Dental Laboratories Process for making a dental restoration model
US9486141B2 (en) * 2011-08-09 2016-11-08 Carestream Health, Inc. Identification of dental caries in live video images
EP2761534B1 (en) 2011-09-28 2020-11-18 FotoNation Limited Systems for encoding light field image files
GB2497517B (en) * 2011-12-06 2016-05-25 Toshiba Res Europe Ltd A reconstruction system and method
US8767040B2 (en) 2012-01-11 2014-07-01 Google Inc. Method and system for displaying panoramic imagery
US10778949B2 (en) 2012-04-06 2020-09-15 Adobe Inc. Robust video-based camera rotation estimation
US9317928B2 (en) 2012-04-06 2016-04-19 Adobe Systems Incorporated Detecting and tracking point features with primary colors
US9786097B2 (en) 2012-06-22 2017-10-10 Matterport, Inc. Multi-modal method for interacting with 3D models
US10163261B2 (en) * 2014-03-19 2018-12-25 Matterport, Inc. Selecting two-dimensional imagery data for display within a three-dimensional model
AU2013305770A1 (en) 2012-08-21 2015-02-26 Pelican Imaging Corporation Systems and methods for parallax detection and correction in images captured using array cameras
US9154805B2 (en) * 2012-09-12 2015-10-06 Advanced Micro Devices, Inc. Video and image compression based on position of the image generating device
US20140100995A1 (en) * 2012-10-05 2014-04-10 Sanu Koshy Collection and Use of Consumer Data Associated with Augmented-Reality Window Shopping
US8948482B2 (en) * 2012-11-01 2015-02-03 Align Technology, Inc. Motion compensation in a three dimensional scan
JP6396310B2 (en) * 2012-11-23 2018-09-26 キャデンス メディカル イメージング インコーポレイテッドCadens Medical Imaging Inc. Method and apparatus for displaying to a user a transition between a first rendering projection and a second rendering projection
US20140172392A1 (en) * 2012-12-18 2014-06-19 Palo Alto Research Center Incorporated Simultaneous mapping and registering thermal images
AU2012268846A1 (en) * 2012-12-21 2014-07-10 Canon Kabushiki Kaisha Optimal patch ranking for coordinate transform estimation of microscope images from sparse patch shift estimates
JP2016509292A (en) 2013-01-03 2016-03-24 メタ カンパニー Extramissive spatial imaging digital eyeglass device or extended intervening vision
US9224368B2 (en) * 2013-02-20 2015-12-29 Google Inc. Merging three-dimensional models of varying resolution
US8866912B2 (en) 2013-03-10 2014-10-21 Pelican Imaging Corporation System and methods for calibration of an array camera using a single captured image
WO2014159779A1 (en) 2013-03-14 2014-10-02 Pelican Imaging Corporation Systems and methods for reducing motion blur in images or video in ultra low light with array cameras
EP2972478B1 (en) * 2013-03-15 2020-12-16 Uatc, Llc Methods, systems, and apparatus for multi-sensory stereo vision for robotics
US9445003B1 (en) * 2013-03-15 2016-09-13 Pelican Imaging Corporation Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information
US9196084B2 (en) * 2013-03-15 2015-11-24 Urc Ventures Inc. Determining object volume from mobile device images
EP2983139A4 (en) * 2013-04-04 2016-12-28 Sony Corp Image processing device, image processing method and program
WO2014161092A1 (en) 2013-04-05 2014-10-09 Cinema Control Laboratories Inc. System and method for controlling an equipment related to image capture
US9479709B2 (en) 2013-10-10 2016-10-25 Nvidia Corporation Method and apparatus for long term image exposure with image stabilization on a mobile device
US10119808B2 (en) 2013-11-18 2018-11-06 Fotonation Limited Systems and methods for estimating depth from projected texture using camera arrays
WO2015081279A1 (en) 2013-11-26 2015-06-04 Pelican Imaging Corporation Array camera configurations incorporating multiple constituent array cameras
EP2884364B1 (en) * 2013-12-12 2018-09-26 Hexagon Technology Center GmbH Autonomous gardening vehicle with camera
US10292624B2 (en) 2014-01-21 2019-05-21 Vorum Research Corporation Method and system for generating a three-dimensional scan of an object
US10010387B2 (en) 2014-02-07 2018-07-03 3Shape A/S Detecting tooth shade
CN103796004B (en) * 2014-02-13 2015-09-30 西安交通大学 A kind of binocular depth cognitive method of initiating structure light
CN104065954B (en) * 2014-07-03 2016-05-18 中国传媒大学 A kind of disparity range method for quick of high definition three-dimensional video-frequency
US9724177B2 (en) 2014-08-19 2017-08-08 Align Technology, Inc. Viewfinder with real-time tracking for intraoral scanning
US11205305B2 (en) 2014-09-22 2021-12-21 Samsung Electronics Company, Ltd. Presentation of three-dimensional video
US10547825B2 (en) 2014-09-22 2020-01-28 Samsung Electronics Company, Ltd. Transmission of three-dimensional video
KR20170063827A (en) 2014-09-29 2017-06-08 포토네이션 케이맨 리미티드 Systems and methods for dynamic calibration of array cameras
KR102264591B1 (en) * 2015-02-27 2021-06-15 삼성전자주식회사 Image Processing Method and Electronic Device supporting the same
EP3265845A4 (en) * 2015-03-05 2019-01-09 Commonwealth Scientific and Industrial Research Organisation Structure modelling
EP3859676A1 (en) 2015-06-30 2021-08-04 Mapillary AB Method in constructing a model of a scenery and device therefor
TWI567364B (en) 2015-09-08 2017-01-21 財團法人工業技術研究院 Structured light generating apparatus, measuring system and method thereof
US20170171525A1 (en) * 2015-12-14 2017-06-15 Sony Corporation Electronic system including image processing unit for reconstructing 3d surfaces and iterative triangulation method
US11195043B2 (en) 2015-12-15 2021-12-07 Cortica, Ltd. System and method for determining common patterns in multimedia content elements based on key points
WO2017105641A1 (en) 2015-12-15 2017-06-22 Cortica, Ltd. Identification of key points in multimedia data elements
US10621433B1 (en) 2015-12-18 2020-04-14 EControls Holdings, KKC Multiscopic whitetail scoring game camera systems and methods
US10037537B2 (en) 2016-02-19 2018-07-31 Alitheon, Inc. Personal history in track and trace system
EP3236401A1 (en) 2016-04-18 2017-10-25 Alitheon, Inc. Authentication-triggered processes
EP3463257A1 (en) 2016-05-27 2019-04-10 3M Innovative Properties Company Coating compositions having hydrophilic and hydrophobic particles for use in intraoral scanning methods
WO2017205144A1 (en) 2016-05-27 2017-11-30 3M Innovative Properties Company Coating compositions having particles with differing refractive indices for use in intraoral scanning methods
US10740767B2 (en) 2016-06-28 2020-08-11 Alitheon, Inc. Centralized databases storing digital fingerprints of objects for collaborative authentication
US10915612B2 (en) 2016-07-05 2021-02-09 Alitheon, Inc. Authenticated production
US10416836B2 (en) * 2016-07-11 2019-09-17 The Boeing Company Viewpoint navigation control for three-dimensional visualization using two-dimensional layouts
US10902540B2 (en) 2016-08-12 2021-01-26 Alitheon, Inc. Event-driven authentication of physical objects
US10839528B2 (en) 2016-08-19 2020-11-17 Alitheon, Inc. Authentication-based tracking
GB2553148A (en) * 2016-08-26 2018-02-28 Nctech Ltd Modelling system and method
TWI651513B (en) 2016-11-15 2019-02-21 財團法人工業技術研究院 Three dimensional measuring system and measuring method thereof
US11003305B2 (en) * 2016-11-18 2021-05-11 Zspace, Inc. 3D user interface
CN106776996B (en) * 2016-12-02 2018-09-07 百度在线网络技术(北京)有限公司 Method and apparatus for the accuracy for testing high-precision map
KR102534875B1 (en) * 2016-12-08 2023-05-22 한국전자통신연구원 Method for generating image of arbitrary view-pont using camera array and multi-focus image and apparatus thereof
EP3340023B1 (en) * 2016-12-22 2020-02-12 Dassault Systèmes Fast manipulation of objects in a three-dimensional scene
WO2018123801A1 (en) * 2016-12-28 2018-07-05 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ Three-dimensional model distribution method, three-dimensional model receiving method, three-dimensional model distribution device, and three-dimensional model receiving device
GB2546459B (en) * 2017-05-10 2018-02-28 Tomlinson Martin Data verification
GB201708520D0 (en) * 2017-05-27 2017-07-12 Dawood Andrew A method for reducing artefact in intra oral scans
WO2019008581A1 (en) 2017-07-05 2019-01-10 Cortica Ltd. Driving policies determination
WO2019012527A1 (en) 2017-07-09 2019-01-17 Cortica Ltd. Deep learning networks orchestration
US10431000B2 (en) * 2017-07-18 2019-10-01 Sony Corporation Robust mesh tracking and fusion by using part-based key frames and priori model
US11062118B2 (en) 2017-07-25 2021-07-13 Alitheon, Inc. Model-based digital fingerprinting
JP7248680B2 (en) 2017-08-10 2023-03-29 ディーフォーディー テクノロジーズ エルエルシー intraoral scanner
WO2019045144A1 (en) * 2017-08-31 2019-03-07 (주)레벨소프트 Medical image processing apparatus and medical image processing method which are for medical navigation device
JP6822929B2 (en) * 2017-09-19 2021-01-27 株式会社東芝 Information processing equipment, image recognition method and image recognition program
EP3698328A1 (en) * 2017-10-19 2020-08-26 InterDigital VC Holdings, Inc. Method and device for predictive encoding/decoding of a point cloud
US11403816B2 (en) * 2017-11-30 2022-08-02 Mitsubishi Electric Corporation Three-dimensional map generation system, three-dimensional map generation method, and computer readable medium
US10469250B2 (en) * 2017-12-22 2019-11-05 Max Adel Rady Physical item mapping to blockchain framework
US11087013B2 (en) 2018-01-22 2021-08-10 Alitheon, Inc. Secure digital fingerprint key object database
US10521970B2 (en) * 2018-02-21 2019-12-31 Adobe Inc. Refining local parameterizations for applying two-dimensional images to three-dimensional models
KR102068489B1 (en) * 2018-03-30 2020-01-22 (주)온넷시스템즈코리아 3d object creation apparatus
CA3096417A1 (en) 2018-04-11 2019-10-17 Align Technology, Inc. Releasable palatal expanders
US11127203B2 (en) * 2018-05-16 2021-09-21 Samsung Electronics Co., Ltd. Leveraging crowdsourced data for localization and mapping within an environment
WO2019243202A1 (en) * 2018-06-21 2019-12-26 3Shape A/S Method, device and system for correlating at least one additional 2d-image to a 3d-representation of at least a part of tooth
US10846544B2 (en) 2018-07-16 2020-11-24 Cartica Ai Ltd. Transportation prediction system and method
US20200068197A1 (en) * 2018-08-27 2020-02-27 Ati Technologies Ulc Benefit-based bitrate distribution for video encoding
US11689707B2 (en) * 2018-09-20 2023-06-27 Shoppertrak Rct Llc Techniques for calibrating a stereoscopic camera in a device
EP3857511A4 (en) * 2018-09-28 2022-05-04 INTEL Corporation Methods and apparatus to generate photo-realistic three-dimensional models of photographed environment
US11126870B2 (en) 2018-10-18 2021-09-21 Cartica Ai Ltd. Method and system for obstacle detection
US10839694B2 (en) 2018-10-18 2020-11-17 Cartica Ai Ltd Blind spot alert
US20200133308A1 (en) 2018-10-18 2020-04-30 Cartica Ai Ltd Vehicle to vehicle (v2v) communication less truck platooning
US11181911B2 (en) 2018-10-18 2021-11-23 Cartica Ai Ltd Control transfer of a vehicle
US11126869B2 (en) 2018-10-26 2021-09-21 Cartica Ai Ltd. Tracking after objects
US10789535B2 (en) 2018-11-26 2020-09-29 Cartica Ai Ltd Detection of road elements
US10963670B2 (en) 2019-02-06 2021-03-30 Alitheon, Inc. Object change detection and measurement using digital fingerprints
US11643005B2 (en) 2019-02-27 2023-05-09 Autobrains Technologies Ltd Adjusting adjustable headlights of a vehicle
US11285963B2 (en) 2019-03-10 2022-03-29 Cartica Ai Ltd. Driver-based prediction of dangerous events
US11694088B2 (en) 2019-03-13 2023-07-04 Cortica Ltd. Method for object detection using knowledge distillation
JP6936826B2 (en) * 2019-03-18 2021-09-22 株式会社モリタ製作所 Image processing equipment, display system, image processing method, and image processing program
US11132548B2 (en) 2019-03-20 2021-09-28 Cortica Ltd. Determining object information that does not explicitly appear in a media unit signature
US10789527B1 (en) 2019-03-31 2020-09-29 Cortica Ltd. Method for object detection using shallow neural networks
US10776669B1 (en) 2019-03-31 2020-09-15 Cortica Ltd. Signature generation and object detection that refer to rare scenes
US11222069B2 (en) 2019-03-31 2022-01-11 Cortica Ltd. Low-power calculation of a signature of a media unit
US10796444B1 (en) 2019-03-31 2020-10-06 Cortica Ltd Configuring spanning elements of a signature generator
US11488290B2 (en) 2019-03-31 2022-11-01 Cortica Ltd. Hybrid representation of a media unit
EP3734506A1 (en) 2019-05-02 2020-11-04 Alitheon, Inc. Automated authentication region localization and capture
EP3736717A1 (en) 2019-05-10 2020-11-11 Alitheon, Inc. Loop chain digital fingerprint method and system
WO2021055585A1 (en) 2019-09-17 2021-03-25 Boston Polarimetrics, Inc. Systems and methods for surface modeling using polarization cues
US11410275B2 (en) * 2019-09-23 2022-08-09 Tencent America LLC Video coding for machine (VCM) based system and method for video super resolution (SR)
BR112022006617A2 (en) 2019-10-07 2022-06-28 Boston Polarimetrics Inc SYSTEMS AND METHODS FOR SENSOR DETECTION OF NORMALS ON THE SURFACE WITH POLARIZATION
US11238146B2 (en) 2019-10-17 2022-02-01 Alitheon, Inc. Securing composite objects using digital fingerprints
EP4066001A4 (en) 2019-11-30 2024-01-24 Boston Polarimetrics Inc Systems and methods for transparent object segmentation using polarization cues
GB201918006D0 (en) * 2019-12-09 2020-01-22 Univ Leeds Innovations Ltd Determining spatial relationship between upper and lower teeth
US11593662B2 (en) 2019-12-12 2023-02-28 Autobrains Technologies Ltd Unsupervised cluster generation
US10748022B1 (en) 2019-12-12 2020-08-18 Cartica Ai Ltd Crowd separation
EP3859603A1 (en) 2020-01-28 2021-08-04 Alitheon, Inc. Depth-based digital fingerprinting
KR20220132620A (en) 2020-01-29 2022-09-30 인트린식 이노베이션 엘엘씨 Systems and methods for characterizing object pose detection and measurement systems
JP2023511747A (en) 2020-01-30 2023-03-22 イントリンジック イノベーション エルエルシー Systems and methods for synthesizing data for training statistical models with different imaging modalities, including polarization imaging
US11590988B2 (en) 2020-03-19 2023-02-28 Autobrains Technologies Ltd Predictive turning assistant
US11341348B2 (en) 2020-03-23 2022-05-24 Alitheon, Inc. Hand biometrics system and method using digital fingerprints
US11568683B2 (en) 2020-03-23 2023-01-31 Alitheon, Inc. Facial biometrics system and method using digital fingerprints
US11538573B2 (en) * 2020-03-30 2022-12-27 James R. Glidewell Dental Ceramics, Inc. Virtual dental restoration insertion verification
US11827215B2 (en) 2020-03-31 2023-11-28 AutoBrains Technologies Ltd. Method for training a driving related object detector
US11948377B2 (en) 2020-04-06 2024-04-02 Alitheon, Inc. Local encoding of intrinsic authentication data
US11663849B1 (en) 2020-04-23 2023-05-30 Alitheon, Inc. Transform pyramiding for fingerprint matching system and method
WO2021243088A1 (en) 2020-05-27 2021-12-02 Boston Polarimetrics, Inc. Multi-aperture polarization optical systems using beam splitters
US11700123B2 (en) 2020-06-17 2023-07-11 Alitheon, Inc. Asset-backed digital security tokens
CN111862305A (en) 2020-06-30 2020-10-30 北京百度网讯科技有限公司 Method, apparatus, and computer storage medium for processing image
US11756424B2 (en) 2020-07-24 2023-09-12 AutoBrains Technologies Ltd. Parking assist
US11445165B1 (en) * 2021-02-19 2022-09-13 Dentsply Sirona Inc. Method, system and computer readable storage media for visualizing a magnified dental treatment site
US11290658B1 (en) 2021-04-15 2022-03-29 Boston Polarimetrics, Inc. Systems and methods for camera exposure control
US11954886B2 (en) 2021-04-15 2024-04-09 Intrinsic Innovation Llc Systems and methods for six-degree of freedom pose estimation of deformable objects
US11689813B2 (en) 2021-07-01 2023-06-27 Intrinsic Innovation Llc Systems and methods for high dynamic range imaging using crossed polarizers

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7085323B2 (en) * 2002-04-03 2006-08-01 Stmicroelectronics, Inc. Enhanced resolution video construction method and apparatus
US20070103460A1 (en) * 2005-11-09 2007-05-10 Tong Zhang Determining camera motion
US20070253618A1 (en) * 2006-03-20 2007-11-01 Samsung Electronics Co., Ltd Camera calibration method and medium and 3D object reconstruction method and medium using the same

Family Cites Families (86)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5276785A (en) * 1990-08-02 1994-01-04 Xerox Corporation Moving viewpoint with respect to a target in a three-dimensional workspace
US5359703A (en) * 1990-08-02 1994-10-25 Xerox Corporation Moving an object in a three-dimensional workspace
US5390108A (en) * 1991-05-24 1995-02-14 Forensic Technology Wai Inc. Computer automated bullet analysis apparatus
FR2700654B1 (en) * 1993-01-19 1995-02-17 Thomson Csf Method for estimating disparity between monoscopic images constituting a stereoscopic image.
US5850352A (en) * 1995-03-31 1998-12-15 The Regents Of The University Of California Immersive video, including video hypermosaicing to generate from multiple video views of a scene a three-dimensional video mosaic from which diverse virtual video scene images are synthesized, including panoramic, scene interactive and stereoscopic images
US5963664A (en) * 1995-06-22 1999-10-05 Sarnoff Corporation Method and system for image combination using a parallax-based technique
US7819807B2 (en) * 1996-06-28 2010-10-26 Sonosite, Inc. Balance body ultrasound system
US6396941B1 (en) * 1996-08-23 2002-05-28 Bacus Research Laboratories, Inc. Method and apparatus for internet, intranet, and local viewing of virtual microscope slides
US6272235B1 (en) * 1997-03-03 2001-08-07 Bacus Research Laboratories, Inc. Method and apparatus for creating a virtual microscope slide
IL119831A (en) * 1996-12-15 2002-12-01 Cognitens Ltd Apparatus and method for 3d surface geometry reconstruction
US6346940B1 (en) * 1997-02-27 2002-02-12 Kabushiki Kaisha Toshiba Virtualized endoscope system
US6208347B1 (en) 1997-06-23 2001-03-27 Real-Time Geometry Corporation System and method for computer modeling of 3D objects and 2D images by mesh constructions that incorporate non-spatial data such as color or texture
US6269175B1 (en) * 1998-08-28 2001-07-31 Sarnoff Corporation Method and apparatus for enhancing regions of aligned images using flow estimation
US6477268B1 (en) * 1998-11-17 2002-11-05 Industrial Technology Research Institute Producing transitions between vistas
US8248457B2 (en) * 1999-02-25 2012-08-21 Visionsense, Ltd. Optical device
US7068825B2 (en) * 1999-03-08 2006-06-27 Orametrix, Inc. Scanning system and calibration method for capturing precise three-dimensional information of objects
US6614452B1 (en) * 1999-11-15 2003-09-02 Xenogen Corporation Graphical user interface for in-vivo imaging
US7234937B2 (en) * 1999-11-30 2007-06-26 Orametrix, Inc. Unified workstation for virtual craniofacial diagnosis, treatment planning and therapeutics
US6525732B1 (en) * 2000-02-17 2003-02-25 Wisconsin Alumni Research Foundation Network-based viewing of images of three-dimensional objects
US6826299B2 (en) * 2000-07-31 2004-11-30 Geodetic Services, Inc. Photogrammetric image correlation and measurement system and method
US7260274B2 (en) * 2000-12-01 2007-08-21 Imax Corporation Techniques and systems for developing high-resolution imagery
US6633317B2 (en) * 2001-01-02 2003-10-14 Microsoft Corporation Image-based walkthrough system and process employing spatial video streaming
US6987512B2 (en) * 2001-03-29 2006-01-17 Microsoft Corporation 3D navigation techniques
JP3842080B2 (en) * 2001-07-06 2006-11-08 グローリー工業株式会社 Signature verification apparatus, signature verification method, and signature verification program
FR2831014B1 (en) * 2001-10-16 2004-02-13 Oreal METHOD AND DEVICE FOR DETERMINING THE DESIRED AND / OR EFFECTIVE DEGREE OF AT LEAST ONE CHARACTERISTIC OF A PRODUCT
US7046840B2 (en) * 2001-11-09 2006-05-16 Arcsoft, Inc. 3-D reconstruction engine
US20030132936A1 (en) * 2001-11-21 2003-07-17 Kevin Kreeger Display of two-dimensional and three-dimensional views during virtual examination
US7184071B2 (en) * 2002-08-23 2007-02-27 University Of Maryland Method of three-dimensional object reconstruction from a video sequence using a generic model
US20040197728A1 (en) 2002-09-10 2004-10-07 Amir Abolfathi Architecture for treating teeth
US7589732B2 (en) * 2002-11-05 2009-09-15 Autodesk, Inc. System and method of integrated spatial and temporal navigation
EP1890261B1 (en) * 2006-08-14 2009-02-18 BrainLAB AG Registration of MR data using generic models
US7695278B2 (en) * 2005-05-20 2010-04-13 Orametrix, Inc. Method and system for finding tooth features on a virtual three-dimensional model
US6968973B2 (en) * 2003-05-31 2005-11-29 Microsoft Corporation System and process for viewing and navigating through an interactive video tour
US7349563B2 (en) * 2003-06-25 2008-03-25 Siemens Medical Solutions Usa, Inc. System and method for polyp visualization
US7990384B2 (en) * 2003-09-15 2011-08-02 At&T Intellectual Property Ii, L.P. Audio-visual selection process for the synthesis of photo-realistic talking-head animations
US7474932B2 (en) * 2003-10-23 2009-01-06 Technest Holdings, Inc. Dental computer-aided design (CAD) methods and systems
US20050089213A1 (en) * 2003-10-23 2005-04-28 Geng Z. J. Method and apparatus for three-dimensional modeling via an image mosaic system
EP1691666B1 (en) * 2003-12-12 2012-05-30 University of Washington Catheterscope 3d guidance and interface system
DE602004017058D1 (en) * 2003-12-18 2008-11-20 Matsushita Electric Ind Co Ltd RECORDING DEVICE, INTEGRATED CIRCUIT FOR A RECORDING DEVICE, RECORDING METHOD, COMPUTER PROGRAM AND COMPUTER RECORDING MEDIUM FOR ENABLEING CONTENT RECORDING
US7693325B2 (en) * 2004-01-14 2010-04-06 Hexagon Metrology, Inc. Transprojection of geometry data
KR100519782B1 (en) * 2004-03-04 2005-10-07 삼성전자주식회사 Method and apparatus for detecting people using a stereo camera
US7502505B2 (en) * 2004-03-15 2009-03-10 Microsoft Corporation High-quality gradient-corrected linear interpolation for demosaicing of color images
WO2006083297A2 (en) 2004-06-10 2006-08-10 Sarnoff Corporation Method and apparatus for aligning video to three-dimensional point clouds
US7015926B2 (en) 2004-06-28 2006-03-21 Microsoft Corporation System and process for generating a two-layer, 3D representation of a scene
US7230620B2 (en) * 2004-08-05 2007-06-12 Mitsubishi Electric Research Laboratories, Inc. Rendering deformable and animated surface reflectance fields
CA2511040A1 (en) 2004-09-23 2006-03-23 The Governors Of The University Of Alberta Method and system for real time image rendering
US20060132482A1 (en) * 2004-11-12 2006-06-22 Oh Byong M Method for inter-scene transitions
US7610560B2 (en) * 2004-11-16 2009-10-27 Microsoft Corporation Methods for automated and semiautomated composition of visual sequences, flows, and flyovers based on content and context
US20060103678A1 (en) * 2004-11-18 2006-05-18 Pascal Cathier Method and system for interactive visualization of locally oriented structures
US7365747B2 (en) * 2004-12-07 2008-04-29 The Boeing Company Methods and systems for controlling an image generator to define, generate, and view geometric images of an object
US20060127852A1 (en) * 2004-12-14 2006-06-15 Huafeng Wen Image based orthodontic treatment viewing system
US7494338B2 (en) * 2005-01-11 2009-02-24 Duane Durbin 3D dental scanner
WO2006085266A1 (en) * 2005-02-08 2006-08-17 Philips Intellectual Property & Standard Gmbh Medical image viewing protocols
US7508430B1 (en) 2005-02-18 2009-03-24 Magnachip Semiconductor, Ltd. Method for locally reducing row noise
US7522755B2 (en) 2005-03-01 2009-04-21 General Electric Company Systems, methods and apparatus for filtered back-projection reconstruction in digital tomosynthesis
US7495582B2 (en) 2005-03-08 2009-02-24 Northrop Grumman Corporation Geographic information storage, transmission and display system
US8228994B2 (en) * 2005-05-20 2012-07-24 Microsoft Corporation Multi-view video coding based on temporal and view decomposition
TW200708091A (en) * 2005-08-03 2007-02-16 Coretronic Corp Projection display apparatus and method thereof
US20070046661A1 (en) * 2005-08-31 2007-03-01 Siemens Medical Solutions Usa, Inc. Three or four-dimensional medical imaging navigation methods and systems
US7978192B2 (en) * 2005-10-04 2011-07-12 Rdv Systems Ltd. Method and apparatus for evaluating sight distance
KR20070039641A (en) 2005-10-10 2007-04-13 주식회사 팬택 Method and apparatus for generating three dimension image using one camera
US8160400B2 (en) 2005-11-17 2012-04-17 Microsoft Corporation Navigating images using image based geometric alignment and object based controls
US7860308B2 (en) * 2005-12-01 2010-12-28 Yahoo! Inc. Approach for near duplicate image detection
US7813591B2 (en) * 2006-01-20 2010-10-12 3M Innovative Properties Company Visual feedback of 3D scan parameters
US7819591B2 (en) 2006-02-13 2010-10-26 3M Innovative Properties Company Monocular three-dimensional imaging
US7372642B2 (en) 2006-02-13 2008-05-13 3M Innovative Properties Company Three-channel camera systems with non-collinear apertures
US20070247454A1 (en) * 2006-04-19 2007-10-25 Norbert Rahn 3D visualization with synchronous X-ray image display
US7764849B2 (en) 2006-07-31 2010-07-27 Microsoft Corporation User interface for navigating through images
US8805034B2 (en) * 2006-08-11 2014-08-12 Koninklijke Philips N.V. Selection of datasets from 3D renderings for viewing
US7581186B2 (en) * 2006-09-11 2009-08-25 Apple Inc. Media manager with integrated browsers
US20080165195A1 (en) * 2007-01-06 2008-07-10 Outland Research, Llc Method, apparatus, and software for animated self-portraits
US9037215B2 (en) * 2007-01-31 2015-05-19 The Penn State Research Foundation Methods and apparatus for 3D route planning through hollow organs
GB0704319D0 (en) * 2007-03-06 2007-04-11 Areograph Ltd Image capture and playback
US7782319B2 (en) * 2007-03-28 2010-08-24 Autodesk, Inc. Three-dimensional orientation indicator and controller
US8467628B2 (en) 2007-04-24 2013-06-18 21 Ct, Inc. Method and system for fast dense stereoscopic ranging
US7843454B1 (en) * 2007-04-25 2010-11-30 Adobe Systems Incorporated Animated preview of images
EP2174301A4 (en) * 2007-06-29 2011-11-16 Three Pixels Wide Pty Ltd Method and system for generating a 3d model from images
US8686991B2 (en) * 2007-09-26 2014-04-01 Autodesk, Inc. Navigation system for a 3D virtual scene
US8451265B2 (en) * 2007-11-16 2013-05-28 Sportvision, Inc. Virtual viewpoint animation
US8073190B2 (en) 2007-11-16 2011-12-06 Sportvision, Inc. 3D textured objects for virtual viewpoint animations
US8503744B2 (en) * 2007-11-19 2013-08-06 Dekel Shlomi Dynamic method and system for representing a three dimensional object navigated from within
US8531449B2 (en) * 2007-12-18 2013-09-10 Navteq B.V. System and method for producing multi-angle views of an object-of-interest from images in an image dataset
DE112009000099T5 (en) 2008-01-04 2010-11-11 3M Innovative Properties Co., St. Paul Image signatures for use in a motion-based three-dimensional reconstruction
US8072448B2 (en) * 2008-01-15 2011-12-06 Google Inc. Three-dimensional annotations for street view data
IL202460A (en) * 2009-12-01 2013-08-29 Rafael Advanced Defense Sys Method and system of generating a three-dimensional view of a real scene
US20110310088A1 (en) * 2010-06-17 2011-12-22 Microsoft Corporation Personalized navigation through virtual 3d environments

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7085323B2 (en) * 2002-04-03 2006-08-01 Stmicroelectronics, Inc. Enhanced resolution video construction method and apparatus
US20070103460A1 (en) * 2005-11-09 2007-05-10 Tong Zhang Determining camera motion
US20070253618A1 (en) * 2006-03-20 2007-11-01 Samsung Electronics Co., Ltd Camera calibration method and medium and 3D object reconstruction method and medium using the same

Cited By (92)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11026768B2 (en) 1998-10-08 2021-06-08 Align Technology, Inc. Dental appliance reinforcement
US11213368B2 (en) 2008-03-25 2022-01-04 Align Technology, Inc. Reconstruction of non-visible part of tooth
US10758321B2 (en) 2008-05-23 2020-09-01 Align Technology, Inc. Smile designer
US10543064B2 (en) 2008-05-23 2020-01-28 Align Technology, Inc. Dental implant positioning
US10842601B2 (en) 2008-06-12 2020-11-24 Align Technology, Inc. Dental appliance
US11471252B2 (en) 2008-10-08 2022-10-18 Align Technology, Inc. Dental positioning appliance having mesh portion
US20150257853A1 (en) 2009-02-02 2015-09-17 Viax Dental Technologies, LLC Dentist tool
US11813127B2 (en) 2009-02-02 2023-11-14 Viax Dental Technologies Llc Tooth restoration system
US10441382B2 (en) 2009-02-02 2019-10-15 Viax Dental Technologies, LLC Dentist tool
US11865653B2 (en) 2009-02-02 2024-01-09 Viax Dental Technologies Llc Method for producing a dentist tool
US11253961B2 (en) 2009-02-02 2022-02-22 Viax Dental Technologies Llc Method for restoring a tooth
US10144100B2 (en) 2009-02-02 2018-12-04 Viax Dental Technologies, LLC Method of preparation for restoring tooth structure
US11083545B2 (en) 2009-03-19 2021-08-10 Align Technology, Inc. Dental wire attachment
US10919209B2 (en) 2009-08-13 2021-02-16 Align Technology, Inc. Method of forming a dental appliance
US11612454B2 (en) 2010-04-30 2023-03-28 Align Technology, Inc. Individualized orthodontic treatment index
US10524881B2 (en) 2010-04-30 2020-01-07 Align Technology, Inc. Patterned dental positioning appliance
US11925517B2 (en) 2011-05-26 2024-03-12 Viax Dental Technologies Llc Dental tool and guidance devices
US11033356B2 (en) 2011-05-26 2021-06-15 Cyrus Tahmasebi Dental tool and guidance devices
US10426572B2 (en) 2011-05-26 2019-10-01 Viax Dental Technologies Llc Dental tool and guidance devices
US10421152B2 (en) 2011-09-21 2019-09-24 Align Technology, Inc. Laser cutting
US10828719B2 (en) 2011-09-21 2020-11-10 Align Technology, Inc. Laser cutting
US11426259B2 (en) 2012-02-02 2022-08-30 Align Technology, Inc. Identifying forces on a tooth
US10893918B2 (en) 2012-03-01 2021-01-19 Align Technology, Inc. Determining a dental treatment difficulty
US10610332B2 (en) 2012-05-22 2020-04-07 Align Technology, Inc. Adjustment of tooth position in a virtual dental model
US10835128B2 (en) 2014-07-07 2020-11-17 Align Technology, Inc. Apparatus for dental confocal imaging
US11369271B2 (en) 2014-07-07 2022-06-28 Align Technology, Inc. Apparatus for dental imaging
US10772506B2 (en) 2014-07-07 2020-09-15 Align Technology, Inc. Apparatus for dental confocal imaging
US10952827B2 (en) 2014-08-15 2021-03-23 Align Technology, Inc. Calibration of an intraoral scanner
US10507088B2 (en) 2014-08-15 2019-12-17 Align Technology, Inc. Imaging apparatus with simplified optical design
US10507089B2 (en) 2014-08-15 2019-12-17 Align Technology, Inc. Imaging apparatus with simplified optical design
US10624720B1 (en) 2014-08-15 2020-04-21 Align Technology, Inc. Imaging apparatus with temperature compensation
US10327872B2 (en) 2014-08-15 2019-06-25 Align Technology, Inc. Field curvature model for confocal imaging apparatus with curved focal surface
US10130445B2 (en) 2014-09-19 2018-11-20 Align Technology, Inc. Arch expanding appliance
US11638629B2 (en) 2014-09-19 2023-05-02 Align Technology, Inc. Arch expanding appliance
US10449016B2 (en) 2014-09-19 2019-10-22 Align Technology, Inc. Arch adjustment appliance
US10537405B2 (en) 2014-11-13 2020-01-21 Align Technology, Inc. Dental appliance with cavity for an unerupted or erupting tooth
WO2016120767A1 (en) * 2015-01-27 2016-08-04 Align Technology, Inc. Method and system for oral-cavity-imaging-and-modeling equipment
US11037466B2 (en) 2015-01-27 2021-06-15 Align Technology, Inc. Training method and system for oral-cavity-imaging-and-modeling equipment
US10504386B2 (en) 2015-01-27 2019-12-10 Align Technology, Inc. Training method and system for oral-cavity-imaging-and-modeling equipment
US11042774B2 (en) 2015-08-20 2021-06-22 Align Technology, Inc. Photograph-based assessment of dental treatments and procedures
US10248883B2 (en) 2015-08-20 2019-04-02 Align Technology, Inc. Photograph-based assessment of dental treatments and procedures
US11554000B2 (en) 2015-11-12 2023-01-17 Align Technology, Inc. Dental attachment formation structure
US11931222B2 (en) 2015-11-12 2024-03-19 Align Technology, Inc. Dental attachment formation structures
US11596502B2 (en) 2015-12-09 2023-03-07 Align Technology, Inc. Dental attachment placement structure
US11103330B2 (en) 2015-12-09 2021-08-31 Align Technology, Inc. Dental attachment placement structure
US10383705B2 (en) 2016-06-17 2019-08-20 Align Technology, Inc. Orthodontic appliance performance monitor
US10888396B2 (en) 2016-06-17 2021-01-12 Align Technology, Inc. Intraoral appliances with proximity and contact sensing
US10470847B2 (en) 2016-06-17 2019-11-12 Align Technology, Inc. Intraoral appliances with sensing
US11612455B2 (en) 2016-06-17 2023-03-28 Align Technology, Inc. Orthodontic appliance performance monitor
US10585958B2 (en) 2016-07-27 2020-03-10 Align Technology, Inc. Intraoral scanner with dental diagnostics capabilities
US10888400B2 (en) 2016-07-27 2021-01-12 Align Technology, Inc. Methods and apparatuses for forming a three-dimensional volumetric model of a subject's teeth
US10509838B2 (en) 2016-07-27 2019-12-17 Align Technology, Inc. Methods and apparatuses for forming a three-dimensional volumetric model of a subject's teeth
US10380212B2 (en) 2016-07-27 2019-08-13 Align Technology, Inc. Methods and apparatuses for forming a three-dimensional volumetric model of a subject's teeth
US10123706B2 (en) 2016-07-27 2018-11-13 Align Technology, Inc. Intraoral scanner with dental diagnostics capabilities
US10507087B2 (en) 2016-07-27 2019-12-17 Align Technology, Inc. Methods and apparatuses for forming a three-dimensional volumetric model of a subject's teeth
US10606911B2 (en) 2016-07-27 2020-03-31 Align Technology, Inc. Intraoral scanner with dental diagnostics capabilities
US10528636B2 (en) 2016-07-27 2020-01-07 Align Technology, Inc. Methods for dental diagnostics
US10595966B2 (en) 2016-11-04 2020-03-24 Align Technology, Inc. Methods and apparatuses for dental images
US11191617B2 (en) 2016-11-04 2021-12-07 Align Technology, Inc. Methods and apparatuses for dental images
US10932885B2 (en) 2016-11-04 2021-03-02 Align Technology, Inc. Methods and apparatuses for dental images
US11273011B2 (en) 2016-12-02 2022-03-15 Align Technology, Inc. Palatal expanders and methods of expanding a palate
US11376101B2 (en) 2016-12-02 2022-07-05 Align Technology, Inc. Force control, stop mechanism, regulating structure of removable arch adjustment appliance
US10993783B2 (en) 2016-12-02 2021-05-04 Align Technology, Inc. Methods and apparatuses for customizing a rapid palatal expander
US11026831B2 (en) 2016-12-02 2021-06-08 Align Technology, Inc. Dental appliance features for speech enhancement
US10548700B2 (en) 2016-12-16 2020-02-04 Align Technology, Inc. Dental appliance etch template
US10456043B2 (en) 2017-01-12 2019-10-29 Align Technology, Inc. Compact confocal dental scanning apparatus
US10918286B2 (en) 2017-01-12 2021-02-16 Align Technology, Inc. Compact confocal dental scanning apparatus
US11712164B2 (en) 2017-01-12 2023-08-01 Align Technology, Inc. Intraoral scanner with moveable opto-mechanical module
US10779718B2 (en) 2017-02-13 2020-09-22 Align Technology, Inc. Cheek retractor and mobile device holder
US11007035B2 (en) 2017-03-16 2021-05-18 Viax Dental Technologies Llc System for preparing teeth for the placement of veneers
US10613515B2 (en) 2017-03-31 2020-04-07 Align Technology, Inc. Orthodontic appliances including at least partially un-erupted teeth and method of forming them
US11045283B2 (en) 2017-06-09 2021-06-29 Align Technology, Inc. Palatal expander with skeletal anchorage devices
US10639134B2 (en) 2017-06-26 2020-05-05 Align Technology, Inc. Biosensor performance indicator for intraoral appliances
US10885521B2 (en) 2017-07-17 2021-01-05 Align Technology, Inc. Method and apparatuses for interactive ordering of dental aligners
US11419702B2 (en) 2017-07-21 2022-08-23 Align Technology, Inc. Palatal contour anchorage
US10842380B2 (en) 2017-07-27 2020-11-24 Align Technology, Inc. Methods and systems for imaging orthodontic aligners
US10517482B2 (en) 2017-07-27 2019-12-31 Align Technology, Inc. Optical coherence tomography for orthodontic aligners
US11633268B2 (en) 2017-07-27 2023-04-25 Align Technology, Inc. Tooth shading, transparency and glazing
US11116605B2 (en) 2017-08-15 2021-09-14 Align Technology, Inc. Buccal corridor assessment and computation
US11123156B2 (en) 2017-08-17 2021-09-21 Align Technology, Inc. Dental appliance compliance monitoring
US10813720B2 (en) 2017-10-05 2020-10-27 Align Technology, Inc. Interproximal reduction templates
US11534268B2 (en) 2017-10-27 2022-12-27 Align Technology, Inc. Alternative bite adjustment structures
US11576752B2 (en) 2017-10-31 2023-02-14 Align Technology, Inc. Dental appliance having selective occlusal loading and controlled intercuspation
US11096763B2 (en) 2017-11-01 2021-08-24 Align Technology, Inc. Automatic treatment planning
US11534974B2 (en) 2017-11-17 2022-12-27 Align Technology, Inc. Customized fabrication of orthodontic retainers based on patient anatomy
US11219506B2 (en) 2017-11-30 2022-01-11 Align Technology, Inc. Sensors for monitoring oral appliances
US11432908B2 (en) 2017-12-15 2022-09-06 Align Technology, Inc. Closed loop adaptive orthodontic treatment methods and apparatuses
US10980613B2 (en) 2017-12-29 2021-04-20 Align Technology, Inc. Augmented reality enhancements for dental practitioners
US11013581B2 (en) 2018-01-26 2021-05-25 Align Technology, Inc. Diagnostic intraoral methods and apparatuses
US10390913B2 (en) 2018-01-26 2019-08-27 Align Technology, Inc. Diagnostic intraoral scanning
US10813727B2 (en) 2018-01-26 2020-10-27 Align Technology, Inc. Diagnostic intraoral tracking
US11937991B2 (en) 2018-03-27 2024-03-26 Align Technology, Inc. Dental attachment placement structure

Also Published As

Publication number Publication date
US8803958B2 (en) 2014-08-12
DE112009000100T5 (en) 2010-11-11
US10503962B2 (en) 2019-12-10
US20110007137A1 (en) 2011-01-13
US20100283781A1 (en) 2010-11-11
US9937022B2 (en) 2018-04-10
DE112009000093T5 (en) 2010-12-09
WO2009089127A1 (en) 2009-07-16
US8830309B2 (en) 2014-09-09
US20180196995A1 (en) 2018-07-12
WO2009089128A1 (en) 2009-07-16
DE112009000101T5 (en) 2010-11-11
WO2009089126A1 (en) 2009-07-16
US11163976B2 (en) 2021-11-02
DE112009000099T5 (en) 2010-11-11
US20110164810A1 (en) 2011-07-07
US20200065558A1 (en) 2020-02-27
US20110043613A1 (en) 2011-02-24
WO2009089125A2 (en) 2009-07-16
DE112009000094T5 (en) 2010-11-11
US20110007138A1 (en) 2011-01-13
WO2009089125A3 (en) 2009-09-24
US9418474B2 (en) 2016-08-16
US8503763B2 (en) 2013-08-06

Similar Documents

Publication Publication Date Title
US8803958B2 (en) Global camera path optimization
US9191648B2 (en) Hybrid stitching
WO2010077380A2 (en) Global camera path optimization
EP2677938B1 (en) Space carving in 3d data acquisition
US7912257B2 (en) Real time display of acquired 3D dental data
EP2620914B1 (en) Three-dimensional scan recovery
CN111784754B (en) Tooth orthodontic method, device, equipment and storage medium based on computer vision
CN115884727A (en) Smart scan for intraoral scanners
Ahmed et al. 3D reconstruction of the human jaw: A new approach and improvements

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09700486

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 12811239

Country of ref document: US

RET De translation (de og part 6b)

Ref document number: 112009000101

Country of ref document: DE

Date of ref document: 20101111

Kind code of ref document: P

122 Ep: pct application non-entry in european phase

Ref document number: 09700486

Country of ref document: EP

Kind code of ref document: A1