US8462208B2 - System and method for tridimensional cartography of a structural surface - Google Patents

System and method for tridimensional cartography of a structural surface Download PDF

Info

Publication number
US8462208B2
US8462208B2 US12/734,034 US73403408A US8462208B2 US 8462208 B2 US8462208 B2 US 8462208B2 US 73403408 A US73403408 A US 73403408A US 8462208 B2 US8462208 B2 US 8462208B2
Authority
US
United States
Prior art keywords
measuring unit
structural surface
module
coordinates
points
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US12/734,034
Other versions
US20100238269A1 (en
Inventor
François Mirallès
Régis Houde
Julien Beaudry
Bruce Hazel
Jean Cöté
Michel Blain
Serge Sarraillon
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hydro Quebec
Original Assignee
Hydro Quebec
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hydro Quebec filed Critical Hydro Quebec
Assigned to HYDRO-QUEBEC reassignment HYDRO-QUEBEC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HOUDE, REGIS, BEAUDRY, JULIEN, BLAIN, MICHEL, COTE, JEAN, HAZEL, BRUCE, SARRAILLON, SERGE, MIRALLES, FRANCOIS
Publication of US20100238269A1 publication Critical patent/US20100238269A1/en
Application granted granted Critical
Publication of US8462208B2 publication Critical patent/US8462208B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/24Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
    • G01B11/25Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/06Systems determining position data of a target
    • G01S17/46Indirect determination of position data
    • G01S17/48Active triangulation systems, i.e. using the transmission and reflection of electromagnetic waves other than radio waves
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging

Definitions

  • the invention relates to a system and a method for tridimensional cartography of a structural surface likely to be submerged.
  • the maintenance of a hydroelectric dam requires the inspection and repair of structures that may be submerged.
  • a simple visual inspection from images or videos may provide qualitative information. This type of inspection allows for example verifying the surface state of the structures or detecting the presence of cracks.
  • quantitative information allows, for example, not only detecting cracks but also accurately locating and gauging them, or establishing the complete survey of a structure to estimate its deformation.
  • a first solution consists in sending divers to take spatial measurements of the structure. The execution of the inspection is then subjected to several constraints. The area must first be physically accessible and this access must be authorized by the safety rules. The environment of a dam indeed piles up the dangers of the underwater environment and the industrial environment. The divers then need enough visibility to be able to take the measurements. The water may be clouded by particles in suspension coming from the vegetation, aquatic organisms or the corrosion of metallic pieces. Furthermore, the quality of the measurements depends on the skill and the experience of the divers. Since the taking of the measurement is manual, the results comprise no notion of confidence or uncertainty.
  • a second solution for the underwater inspection resides in the use of a robotic system.
  • it may be any remote-controlled or autonomous system allowing collecting the data required for the inspection.
  • the mobile robots such as the ROV (“Remotely Operated Vehicle”) and the AUV (“Autonomous Underwater Vehicle”).
  • ROV Remotely Operated Vehicle
  • AUV Automatic Underwater Vehicle
  • the use of a robot provides numerous advantages and new capabilities with respect to an inspection with divers.
  • the safety requirements are not the same since no operator is directly present in the area to be inspected.
  • the data may be collected by a larger number of various sensors: camera, video camera but also sonar, passive or active vision system. It is possible to combine the use of these sensors to take the measurements in different conditions.
  • the use of a vision system provides a good accuracy for a local inspection but an acoustic system allows collecting data over a larger range and in bad visibility conditions.
  • the filtering and merging of the measurements coming from several sensors may reduce their uncertainty.
  • the amount of collected data is greater.
  • the risks of not detecting an irregularity are lower.
  • a greater acquisition capacity makes large scale structure inspection also possible. It is for these reasons that the underwater robots are actively developed since over thirty years. The maturity of the technology now makes their use possible in the industry.
  • the robotized inspection systems comprise numerous advantages but their tuning poses new problems and their functioning may fail in certain specific conditions.
  • each sensor is efficient for a particular purpose.
  • One of the common problems of the underwater robots is the accurate positional tracking of the system. Yet, to be usable, the collected data need to be referenced in a same global reference system. However, on the outskirts of a dam, a robot sometimes operates in a closed environment, which may bring its position tracking system to a fault. For example, in the case of acoustic systems, the emitted waves bounce back on the walls and adversely affect the interpretation of the reflected signal. Or in the case of compasses, the earth magnetic field is distorted by the presence of the reinforcements of the concrete and prevents a good reading of the orientation.
  • a confidence level in the measurement is also required. This confidence level indicates the reliability of the system and depends of both the position tracking system and the sensor that performs the measurement.
  • An inspection system intended for on-site operation is subjected to perturbations commonly found in the environment (particles in suspension, vibrations, impacts). Thus, even if the system has a good measurement accuracy in ideal conditions, this accuracy will degrade in real conditions.
  • An object of the present invention is to provide a system and a method allowing mapping a structural surface likely to be submerged, which simplify the tracking of the measuring unit with respect to the surface.
  • Another object of the present invention is to provide such a system and a such a method which allow establishing a depth map of the structural surface with a level of accuracy that may reach a fraction of a millimeter, for example 0.1 mm.
  • Another object of the present invention is to provide such a system and such a method capable of determining if a measurement is valid and able to associate a representation of uncertainty to a measurement if desired.
  • Another object of the present invention is to provide such a system having a robustness allowing using it in an underwater environment in possibly harsh conditions, and a reduced sensitivity to perturbations such as particles in suspension, vibrations, impacts.
  • Another object of the present invention is to provide such a system having a certain autonomy of operation.
  • Another object of the present invention is to provide such a system and such a method having a degree of versatility such that they are rather easily usable or practicable, and may self-calibrate.
  • Another object of the present invention is to provide such a system and such a method that allow fast taking of measurements.
  • a system for tridimensional cartography of a structural surface comprising:
  • a method for tridimensional cartography of a structural surface comprising the steps of:
  • FIG. 1 is a schematic diagram illustrating a system for producing cartographic data of a structural surface according to the invention.
  • FIG. 2 is a schematic diagram illustrating a support assembly of the system.
  • FIG. 3 is a schematic diagram illustrating a measuring unit of the system.
  • FIG. 4 is a schematic diagram illustrating a top view of a measuring unit of the system.
  • FIG. 5 is a schematic diagram illustrating an alternate configuration of a measuring unit of the system.
  • FIG. 6 is a schematic diagram illustrating an image captured by the measuring unit.
  • FIG. 7 is a schematic diagram illustrating an acquisition chain of the system.
  • FIG. 8 is a schematic diagram illustrating a system mounted between rails of a dam gate.
  • FIG. 9 is a schematic diagram illustrating a displacement system for a measuring unit of the system.
  • FIG. 1 there is shown a diagram of the system for tridimensional cartography of a structural surface 2 (shown in FIG. 3 ) according to the invention.
  • the system comprises a measuring unit 4 , a support assembly 6 configured to move the measuring unit 4 , and a processing unit 8 connected to the measuring unit 4 and to the support assembly 6 .
  • the system also comprises two wires 10 , 12 in parallel and extendable along the structural surface 2 while being spaced from each other so as to define a reference surface in front of the structural surface 2 .
  • the measuring unit 4 comprises a laser arrangement formed of two lasers 14 , 16 configured to project two distinct light planes 18 , 20 directed toward the surface 2 , and a camera 22 provided with a lens 21 .
  • the wires 10 , 12 form a reference surface that is not planar or in the case where the wires 10 , 12 are not in parallel
  • the out-of-parallelism or flatness defect relative to the measuring unit 4 will be used to bring the desired corrections to the measurements performed by the processing unit 8 (shown in FIG. 1 ).
  • the camera 22 is configured to capture images 28 of the surface 2 containing lines 24 , 26 resulting from an intersection of the light planes 18 , 20 with the surface 2 and four reference points P M1 , P M2 , P M3 , P M4 resulting from an intersection of the light planes 18 , 20 with the wires 10 , 12 .
  • the support assembly 6 carrying the measuring unit 4 may take the form of a frame comprising a mobile element 36 such as an arm to which the measuring unit 4 is attached, and a set of transverse and vertical guiding elements 30 , 32 , 34 for the mobile element 36 .
  • the support assembly 6 is thus capable of controllably moving the measuring unit 4 in front of the surface 2 (shown in FIG. 3 ) so as to progressively scan the surface 2 (in part or completely).
  • the mobile assembly may take various forms, provided that its construction allows reaching the surface 2 to be inspected and controllably moving the measuring unit 4 in front of the surface 2 to build up the depth map of the surface 2 .
  • the transverse guiding element 30 may be made of a rigid lattice as shown in FIG. 8 , slideably mounted on lateral guides 32 , 34 forming the vertical guiding elements, to allow moving the lattice 30 vertically (or in an equivalent direction depending on the situation/orientation of the surface to be examined) with respect to the surface 2 as depicted by the arrows 38 , 40 (shown in FIG. 2 ).
  • the arm 36 is slideably mounted on the lattice 30 to allow moving it horizontally (or in an equivalent direction depending on the situation/orientation of the surface to be examined) with respect to the surface 2 as depicted by the arrows 42 , 44 (shown in FIG. 2 ).
  • the sliding of the arm 36 with respect to the lattice 30 (shown in FIG. 8 ) to position and move the measuring unit 4 (shown in FIG. 8 ) may be achieved by means of gearings driven by one or several motors mounted or not on the frame 6 depending on the type of motor—pneumatic, electric, hydraulic, etc.
  • a submersible electric rotary motor 112 provided with a pinion 111 engaging a rack 110 may be used to move the arm 36 along a rail arrangement 60 .
  • the mobile element 36 will provide a precise positioning to the measuring unit 4 .
  • a submersible rotary optical encoder 114 provided with a pinion 115 engaging the rack 110 may be used to measure the horizontal displacement of the mobile element 36 .
  • a proportional control unit may then control the position of the mobile element 36 .
  • a submersible inclinometer 118 measuring the inclination according to two degrees of liberty with respect to the horizontal may be used.
  • the lasers 14 , 16 of the measuring unit 4 are rigidly secured to the camera 22 .
  • Mirrors 46 , 48 may be used to redirect the light planes 18 , 20 toward the surface 2 depending on the chosen setup.
  • the lasers 14 , 16 may be disposed on a same side of the camera 22 .
  • Such an arrangement, shown in FIG. 5 allows reducing the size of the measuring unit 4 .
  • a single laser may also be used instead of two. In such a case, an appropriate optical arrangement (not shown) may be used to produce the two light planes 14 , 16 .
  • the light planes 14 , 16 may be produced one after the other, so as to appear in subsequent images captured by the camera 22 and then processed in order to take this factor into account.
  • FIG. 5 shows an arrangement of the measuring unit 4 where the laser projectors 14 , 16 respectively project the light planes 17 , 19 and 18 , 20 , lines 23 , 25 and 24 , 26 then resulting from the intersection of the light planes 17 , 19 and 18 , 20 with the surface 2 .
  • the wires 10 , 12 may be formed of steel wires or other types of wires if desired.
  • wires 10 , 11 , 12 disposed in front of the surface 2 , so that the light planes 24 , 26 cross them.
  • the extra wire or wires may be in the same plane as the wires 10 , 12 and parallel to the wires 10 , 12 , or they may be in another plane, preferably parallel to the plane formed by the wires 10 , 12 , and still parallel to the wires 10 , 12 .
  • the addition of extra wires provides more information for the position tracking of the measuring unit 4 and the on-line calibration of its parameters.
  • At least four points P M1 , P M2 , P M3 , P M4 on the wires 10 , 12 are lit by the laser arrangement 14 , 16 .
  • An extra point on the wire 11 located at a distance g, which is not necessarily known, from the reference surface formed by the wires 10 , 12 allows on-line calibrating the set of parameters of each light planes 18 , 20 .
  • the processing unit 8 may be embodied so as to build up a depth map of the surface 2 by means of an iterative and feedback process.
  • the processing unit 8 comprises a tracking module 81 , a mapping module 82 and a decision module 83 .
  • the tracking module 81 determines the location of the measuring unit 4 with respect to a fixed global reference system.
  • the mapping module 82 determines the 3D coordinates of points on the surface 2 in the reference system bound to the camera 22 (shown in FIG. 3 ) of the measuring unit 4 .
  • the decision module 83 determines the commands to transmit to the support 6 that moves the measuring unit 4 in order to efficiently collect the data.
  • the tracking module 81 comprises two modeling sub-modules 56 , 58 , an estimation sub-module 812 , a P Mi points detection sub-module 810 and a distortions correction sub-module 811 .
  • the modeling sub-module 56 models the support assembly 6 by means of a set of parametric data that characterize it.
  • the modeling sub-module 58 models the measuring unit 4 by means of a set of parametric data that characterize it.
  • the P Mi points detection sub-module 810 determines the image coordinates of the points P M1 , P M2 , P M3 , P M4 .
  • the distortions correction sub-module 811 corrects the image coordinates of the detected points in order to cancel the effect of the distortions of the lens 21 of the camera 22 or to take into account an out-of-parallelism defect of the wires 10 , 12 (shown in FIG. 3 ).
  • the estimation sub-module 812 estimates the pose (i.e. the position and orientation) of the measuring unit 4 with respect to the global reference system based on the signals provided by the modeling sub-module 58 and the distortions correction sub-module 811 , and corrects the internal parameters of the measuring unit 4 through a feedback loop.
  • the estimation sub-module 812 may use a Kalman filter to perform the desired estimation.
  • the parametric data of the measuring unit 4 comprise the internal and external parameters of the camera 22 , the parameters of the light planes 18 , 20 and the spacing of the wires 10 , 12 .
  • the typical external parameters of the camera 22 are the orientation and the position with respect to the reference system bound to the wires 10 , 12 .
  • the parameter that defines the spacing of the wires 10 , 12 is marked 2 h .
  • the parametric data of the support assembly 6 typically comprise the position and the orientation with respect to a fixed global reference system as well as the position of the measuring unit 4 on the transverse rail arrangement 60 (shown in FIG. 9 ).
  • the mapping module 82 comprises a P u points detection sub-module 820 that determines the image coordinates of the points contained in the lines 24 , 26 , a distortions correction sub-module 811 similar to that of the tracking module 81 , and a triangulation sub-module 822 .
  • the function of the triangulation sub-module 822 is to compute the 3D coordinates of the points P u in the reference system bound to the camera 22 of the measuring unit 4 .
  • the mapping module 82 also comprises a reference change sub-module 840 that transforms the coordinates of the points P u from the reference system bound to the camera toward the global reference system, and a surface modeling sub-module 54 that progressively builds up the map of the surface 2 .
  • a reference change sub-module 840 that transforms the coordinates of the points P u from the reference system bound to the camera toward the global reference system
  • a surface modeling sub-module 54 that progressively builds up the map of the surface 2 .
  • the surface modeling sub-module 54 may contain a set of 3D points that form the map of the surface 2 and define its tridimensional cartography.
  • the decision module 83 comprises a decision taking sub-module 830 that transmits positioning commands to the support assembly 6 and a communication interface 832 allowing initializing the processing unit 8 and retrieving the information sought, namely the map of the surface 2 , the position and orientation data of the support assembly 6 , and the parametric data of the measuring unit 4 .
  • the measuring unit 4 of the system may be located with a great precision.
  • the system also maps at the same time at least one line of points of the surface 2 by the use of a single image.
  • the camera 22 transmits the images 28 to the processing unit 8 .
  • the processing unit 8 processes the images 28 so as to locate the measuring unit 4 with respect to the reference system bound to the wires 10 , 12 and so as to determine the coordinates of 3D points of the surface 2 .
  • the image coordinates of the points P M1 , P M2 , P M3 , and P M4 are used to locate the measuring unit 4 while the lines 24 , 26 are used to map the surface 2 .
  • the system according to the invention thus allows mapping surfaces of great length while preserving a high precision.
  • the decision sub-module 83 initially transmits positioning commands to the support assembly 6 in order to position the measuring unit 4 at the level of the surface 2 to be inspected.
  • the decision sub-module 83 locks the displacement of the support assembly 6 .
  • the estimation sub-module 812 e.g. Kalman filter
  • the decision sub-module 83 transmits positioning commands to the support assembly 6 to move the measuring unit 4 crosswise (in the directions of arrows 42 , 44 shown in FIG.
  • the mapping module 82 takes measurements at specified intervals via the communication interface 832 . For each pair of lines 24 , 26 , an image is transmitted by the camera 22 to the mapping module 82 .
  • the surface modeling sub-module 54 builds up the map of the surface 2 line by line. The obtained map covers a strip of the surface 2 having a certain height. If several strips are required to map the whole or a desired part of the surface 2 , then the decision module 83 transmits positioning commands to the support assembly 6 in order to move the measuring unit 4 heightwise toward the next adjoining area and then the above process is repeated.
  • the displacement of the measuring unit 4 may be made in a continuous mode or step by step.
  • a continuous displacement a single measurement for a given point of the surface 2 is taken by the tracking module 81 and by the mapping module 82 .
  • several measurements may be taken at each stop and may be combined so as to increase the confidence in the measurements.
  • the mapping system is likely to be subjected to multiple perturbations. Although it might be possible to directly take measurements of the structural surface 2 and then to reference these measurements using the position data of the measuring unit 4 and of the support assembly 6 , these measurements are imprecise, or even false, due to the perturbations met.
  • the system takes advantage of the wires 10 , 12 through an iterative and feedback process, implemented in the tracking module 81 , which ensures the validity of the position data of the measuring unit 4 .
  • This iterative and feedback process uses the model of the measuring unit 4 defined in the sub-module 58 .
  • the minimal set of data for computing the pose of a camera without ambiguity is with four coplanar points.
  • the P M1 points detector 810 determines the image coordinates of the points P M1 , P M2 , P M3 , and P M4 .
  • the determination of the coordinates of the points P M1 , P M2 , P M3 , and P M4 may be achieved using computer vision library algorithms.
  • the distortions corrector 811 corrects these coordinates to cancel the effect of the possible distortions of the lens 21 of the camera 22 .
  • the points P M1 , P M2 , P M3 , and P M4 are then triangulated in the reference system bound to the camera 22 in order to obtain their 3D coordinates. Equation 1 may be used for this purpose:
  • [ X Y Z ] CAM [ ( u - u 0 ) ⁇ Z / ⁇ u ( v - v 0 ) ⁇ Z / ⁇ v - d / ( a + b ⁇ ( u - u 0 ) / ⁇ u + c ⁇ ( v - v 0 ) / ⁇ v ) ] CAM ( 1 ) where (u, v) represent the image coordinates of a point P, and (X, Y, Z) represent the 3D coordinates of P in the reference system bound to the camera 22 .
  • the orientation of the camera may be computed using the algorithm proposed by B. K. P.
  • the position/orientation of the camera 22 thus computed comes as initial parameter of the Kalman filter 812 that refines the estimate of this position/orientation based on the parametric data of the model of the measuring unit 4 provided by the sub-module 58 , the model of the support assembly 6 provided by the sub-module 56 , and the computed pose.
  • the Kalman filter 812 After several iterations of the Kalman filter 812 , if the difference between the coordinates of the points P M1 , P M2 , P M3 , et P M4 estimated by the model of the measuring unit 4 and the coordinates of the points P M1 , P M2 , P M3 , and P M4 detected by the P Mi points detector 810 is above a certain threshold, then the pose is rejected.
  • this technique allows having a more robust and uncorrupted system. It is made possible by the presence of the wires 10 , 12 . This technique thus allows detecting if the wires 10 , 12 are moving in response to perturbations or also if the parametric data of the model of the measuring unit 4 are false.
  • the third wire 11 (shown in FIG. 5 ) is used. These parameters are estimated by the Kalman filter 812 which has the position and the orientation of the camera 22 and the parameters a, b, c, d of the laser planes 18 , 20 as state vector.
  • the P Mi points detector 810 determines the image coordinates of the points P M1 , P M2 , P M3 , P M4 , P M5 , P M6 that are transmitted to the distortions corrector 811 and then to the Kalman filter 812 .
  • the position of the third wire 11 with respect to the wires 10 , 12 is not necessarily known and may also be estimated by the Kalman filter 812 .
  • the P Li points detector 820 determines the 2D coordinates of points of the images 28 contained in the lines 24 , 26 . This processing step may be achieved using computer vision library algorithms.
  • the distortions corrector 811 then corrects the coordinates of the points P Li so as to cancel the effect of possible distortions of the lens 21 of the camera 22 .
  • the 3D coordinates of the points P Li expressed in the reference system bound to the measuring unit 4 , are determined from a single measurement by the triangulation sub-module 822 which uses the equation 1.
  • a change of reference system 840 is then achieved to obtain the coordinates of P u in the global reference system.
  • the reference change may be achieved following the equation:
  • GLOBAL T ⁇ ( GLOBAL , CAM ) ⁇ [ X Y Z 1 ] CAM ( 3 )
  • T is a homogenous transformation matrix that is function of the orientation and of the position of the camera 22 with respect to the global reference system.
  • the positioning of the measuring unit 4 on the support assembly 6 may be performed by the decision taking sub-module 830 in order to be able to control the positioning as function of the processed measurements. For example, it may be possible that the system encounters perturbations (for example, motions of the wires 10 , 12 ) and is no longer able to take measurements.
  • the decision taking sub-module 830 may in that case stop the displacement of the measuring unit 4 until the perturbations disappear.
  • the positioning may also be pre-programmed so as to be achieved without feedback. Or an operator may take control of the displacements via the communication interface 832 .
  • the lines 24 , 26 may be indexed by means of an optical encoder 114 (shown in FIG. 9 ) that provides the position of the measuring unit 4 on the transverse rail arrangement 60 of the support assembly 6 (shown in FIG. 2 ).
  • the index of the lines 24 , 26 corresponds to the position of the measuring unit 4 on the rail arrangement 60 of the support assembly 6 .
  • the configuration of the tracking, mapping and decision modules 81 , 82 , 83 may be adapted for example to refine certain processings in order to produce the mapping data or according to the operating parameters and the type of equipment forming the measuring unit 4 and the support assembly 6 .
  • the processing unit 8 and the communication interface 832 may advantageously be implemented in a processing device 62 such as a PC (computer) or appropriate dedicated circuits controlling the functions of the system.
  • the processing device 62 is connected to the measuring unit 4 and retrieves the images 28 from the camera 22 (shown in FIG. 2 ) and processes them to locate the measuring unit 4 and map the surface 2 .
  • the processing may be achieved in continuous mode at the refresh rate of the camera 22 .
  • the processing device 62 may also control the two displacement axes of the support assembly 6 .
  • the processing device 62 retrieves the position (X, Y) of the measuring unit 4 and stores a profile line in memory. This information is used to express the coordinates of the laser lines in a global reference system bound to the scene.
  • the camera 22 is preferably provided with a high resolution progressive CCD matrix.
  • a camera may communicate with the processing device 62 (shown in FIG. 7 ) via a digital transmission channel such as IEEE-1394 (also known under the denomination FireWire), Camera Link, USB or Gigabyte Ethernet.
  • IEEE-1394 also known under the denomination FireWire
  • the laser projectors 14 , 16 preferably emit at a wavelength in the visible range (between 400 nm and 800 nm).
  • the lens 21 of the camera 22 may use an optical filter that let only the wavelength of the laser projectors 14 , 16 through.
  • the system is designed to operate underwater, it may also be used outside of water if desired. In such a case, the calibration of the measuring unit 4 is simply made outside of water, and the system then works in the same way as underwater.

Abstract

System and method for tridimensional cartography of a structural surface. Two wires are extended in front and along the structural surface so as to define a reference surface. A measuring unit comprising a laser arrangement and a camera is moved in front of the structural surface so as to progressively scan the surface. Tow distinct light planes directed toward the structural surface are projected by means of the laser arrangement. Images of the structural surface containing lines resulting from an intersection of the light planes with the structural surface and four reference points resulting from an intersection of the light planes with the wires are captured by means of the camera. The images are processed to determine the 3D coordinates of the lines defining the mapping in a reference system bound to the reference surface considering the position and the orientation of the measuring unit based on the reference points.

Description

FIELD OF THE INVENTION
The invention relates to a system and a method for tridimensional cartography of a structural surface likely to be submerged.
BACKGROUND
The maintenance of a hydroelectric dam requires the inspection and repair of structures that may be submerged. A simple visual inspection from images or videos may provide qualitative information. This type of inspection allows for example verifying the surface state of the structures or detecting the presence of cracks. However, to establish a more workable evaluation of the state of the installations, it is necessary to have quantitative information. Such information allows, for example, not only detecting cracks but also accurately locating and gauging them, or establishing the complete survey of a structure to estimate its deformation.
The inspection of an underwater structure may be achieved in different ways. A first solution consists in sending divers to take spatial measurements of the structure. The execution of the inspection is then subjected to several constraints. The area must first be physically accessible and this access must be authorized by the safety rules. The environment of a dam indeed piles up the dangers of the underwater environment and the industrial environment. The divers then need enough visibility to be able to take the measurements. The water may be clouded by particles in suspension coming from the vegetation, aquatic organisms or the corrosion of metallic pieces. Furthermore, the quality of the measurements depends on the skill and the experience of the divers. Since the taking of the measurement is manual, the results comprise no notion of confidence or uncertainty. Finally, the time needed for the taking of a measurement may be long whereas the stopping of a part of a dam during the inspection may be very expensive for the operator. All of these constraints result in that the installations that require it are not always regularly inspected. In that case, the state of the dam deteriorates until the day one of its parts malfunctions. To sum up, an inspection with divers is not always possible, has a high cost, exhibits dangers for the human life and limits a rigorous quantitative use of the results.
A second solution for the underwater inspection resides in the use of a robotic system. In general, it may be any remote-controlled or autonomous system allowing collecting the data required for the inspection. In that class are found the mobile robots such as the ROV (“Remotely Operated Vehicle”) and the AUV (“Autonomous Underwater Vehicle”). The use of a robot provides numerous advantages and new capabilities with respect to an inspection with divers. The safety requirements are not the same since no operator is directly present in the area to be inspected. The data may be collected by a larger number of various sensors: camera, video camera but also sonar, passive or active vision system. It is possible to combine the use of these sensors to take the measurements in different conditions. For example, the use of a vision system provides a good accuracy for a local inspection but an acoustic system allows collecting data over a larger range and in bad visibility conditions. The filtering and merging of the measurements coming from several sensors may reduce their uncertainty. The amount of collected data is greater. By having a sufficient measurement density, the risks of not detecting an irregularity are lower. A greater acquisition capacity makes large scale structure inspection also possible. It is for these reasons that the underwater robots are actively developed since over thirty years. The maturity of the technology now makes their use possible in the industry.
The robotized inspection systems comprise numerous advantages but their tuning poses new problems and their functioning may fail in certain specific conditions.
Although the data can be collected with different types of sensors, each sensor is efficient for a particular purpose.
One of the common problems of the underwater robots is the accurate positional tracking of the system. Yet, to be usable, the collected data need to be referenced in a same global reference system. However, on the outskirts of a dam, a robot sometimes operates in a closed environment, which may bring its position tracking system to a fault. For example, in the case of acoustic systems, the emitted waves bounce back on the walls and adversely affect the interpretation of the reflected signal. Or in the case of compasses, the earth magnetic field is distorted by the presence of the reinforcements of the concrete and prevents a good reading of the orientation.
The inspection of certain structures requires high measurement accuracy. Therefore, it is both necessary that the sensor that effectively performs the measurement be accurate and that the position tracking system of the robot be also accurate. Currently, a data measurement accuracy expressed in a global reference system in the order of the centimeter is generally reached, which is inadequate for certain applications.
It should be noted that a high measurement accuracy is not enough. A confidence level in the measurement is also required. This confidence level indicates the reliability of the system and depends of both the position tracking system and the sensor that performs the measurement.
An inspection system intended for on-site operation is subjected to perturbations commonly found in the environment (particles in suspension, vibrations, impacts). Thus, even if the system has a good measurement accuracy in ideal conditions, this accuracy will degrade in real conditions.
SUMMARY
An object of the present invention is to provide a system and a method allowing mapping a structural surface likely to be submerged, which simplify the tracking of the measuring unit with respect to the surface.
Another object of the present invention is to provide such a system and a such a method which allow establishing a depth map of the structural surface with a level of accuracy that may reach a fraction of a millimeter, for example 0.1 mm.
Another object of the present invention is to provide such a system and such a method capable of determining if a measurement is valid and able to associate a representation of uncertainty to a measurement if desired.
Another object of the present invention is to provide such a system having a robustness allowing using it in an underwater environment in possibly harsh conditions, and a reduced sensitivity to perturbations such as particles in suspension, vibrations, impacts.
Another object of the present invention is to provide such a system having a certain autonomy of operation.
Another object of the present invention is to provide such a system and such a method having a degree of versatility such that they are rather easily usable or practicable, and may self-calibrate.
Another object of the present invention is to provide such a system and such a method that allow fast taking of measurements.
According to an aspect of the present invention, there is provided a system for tridimensional cartography of a structural surface, comprising:
    • two wires extendable along the structural surface while being spaced from each other so as to define a reference surface in front of the structural surface;
    • a measuring unit comprising a laser arrangement configured to project two distinct light planes directed toward the structural surface and crossing the wires, and a camera configured to capture images of the structural surface containing lines resulting from an intersection of the light planes with the structural surface and reference points resulting from an intersection of the light planes with the wires;
    • a support assembly configured to controllably move the measuring unit in front of the structural surface so as to progressively scan the structural surface to be mapped; and
    • a processing unit for processing the images captured by the camera, the processing unit having:
      • a processing circuit that determines an orientation and a position of the measuring unit with respect to the reference surface as function of the reference points in the images;
      • a processing circuit that determines 3D coordinates of points defining the lines in the images with respect to a reference system bound to the measuring unit determined as function of the orientation and the position of the measuring unit; and
      • a processing circuit that transforms the 3D coordinates into cartographic data representative of the structural surface in a global reference system bound to the reference surface defined by the wires.
According to another aspect of the present invention, there is provided a method for tridimensional cartography of a structural surface, comprising the steps of:
    • a) extending two wires along the structural surface while spacing them from each other so as to define a reference surface in front of the structural surface;
    • b) controllably moving a measuring unit comprising a laser arrangement and a camera in front of the structural surface so as to progressively scan the structural surface;
    • c) projecting two distinct light planes directed toward the structural surface using the laser arrangement;
    • d) capturing images of the structural surface containing lines resulting from an intersection of the light planes with the structural surface and four reference points resulting from an intersection of the light planes with the wire using the camera;
    • e) determining an orientation and a position of the measuring unit with respect to the reference surface as function of the reference points in the images;
    • f) determining 3D coordinates of points defining the lines in the images with respect to a reference system bound to the measuring unit determined as function of the orientation and the position of the measuring unit; and
    • g) transforming the 3D coordinates into cartographic data representative of the structural surface in a reference system bound to the reference surface defined by the wires.
BRIEF DESCRIPTION OF THE DRAWINGS
A detailed description of the preferred embodiments of the invention will be given hereinbelow with reference to the following drawings, in which like numbers refer to like elements:
FIG. 1 is a schematic diagram illustrating a system for producing cartographic data of a structural surface according to the invention.
FIG. 2 is a schematic diagram illustrating a support assembly of the system.
FIG. 3 is a schematic diagram illustrating a measuring unit of the system.
FIG. 4 is a schematic diagram illustrating a top view of a measuring unit of the system.
FIG. 5 is a schematic diagram illustrating an alternate configuration of a measuring unit of the system.
FIG. 6 is a schematic diagram illustrating an image captured by the measuring unit.
FIG. 7 is a schematic diagram illustrating an acquisition chain of the system.
FIG. 8 is a schematic diagram illustrating a system mounted between rails of a dam gate.
FIG. 9 is a schematic diagram illustrating a displacement system for a measuring unit of the system.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
Referring to FIG. 1, there is shown a diagram of the system for tridimensional cartography of a structural surface 2 (shown in FIG. 3) according to the invention. The system comprises a measuring unit 4, a support assembly 6 configured to move the measuring unit 4, and a processing unit 8 connected to the measuring unit 4 and to the support assembly 6.
Referring to FIG. 3, the system also comprises two wires 10, 12 in parallel and extendable along the structural surface 2 while being spaced from each other so as to define a reference surface in front of the structural surface 2. The measuring unit 4 comprises a laser arrangement formed of two lasers 14, 16 configured to project two distinct light planes 18, 20 directed toward the surface 2, and a camera 22 provided with a lens 21. In the case where the wires 10, 12 form a reference surface that is not planar or in the case where the wires 10, 12 are not in parallel, the out-of-parallelism or flatness defect relative to the measuring unit 4 will be used to bring the desired corrections to the measurements performed by the processing unit 8 (shown in FIG. 1).
Referring to FIGS. 3 and 6, the camera 22 is configured to capture images 28 of the surface 2 containing lines 24, 26 resulting from an intersection of the light planes 18, 20 with the surface 2 and four reference points PM1, PM2, PM3, PM4 resulting from an intersection of the light planes 18, 20 with the wires 10, 12.
Referring to FIGS. 2 and 8, the support assembly 6 carrying the measuring unit 4 may take the form of a frame comprising a mobile element 36 such as an arm to which the measuring unit 4 is attached, and a set of transverse and vertical guiding elements 30, 32, 34 for the mobile element 36. The support assembly 6 is thus capable of controllably moving the measuring unit 4 in front of the surface 2 (shown in FIG. 3) so as to progressively scan the surface 2 (in part or completely). The mobile assembly may take various forms, provided that its construction allows reaching the surface 2 to be inspected and controllably moving the measuring unit 4 in front of the surface 2 to build up the depth map of the surface 2. For example, the transverse guiding element 30 may be made of a rigid lattice as shown in FIG. 8, slideably mounted on lateral guides 32, 34 forming the vertical guiding elements, to allow moving the lattice 30 vertically (or in an equivalent direction depending on the situation/orientation of the surface to be examined) with respect to the surface 2 as depicted by the arrows 38, 40 (shown in FIG. 2). The arm 36 is slideably mounted on the lattice 30 to allow moving it horizontally (or in an equivalent direction depending on the situation/orientation of the surface to be examined) with respect to the surface 2 as depicted by the arrows 42, 44 (shown in FIG. 2).
Referring to FIG. 9, the sliding of the arm 36 with respect to the lattice 30 (shown in FIG. 8) to position and move the measuring unit 4 (shown in FIG. 8) may be achieved by means of gearings driven by one or several motors mounted or not on the frame 6 depending on the type of motor—pneumatic, electric, hydraulic, etc. For example, a submersible electric rotary motor 112 provided with a pinion 111 engaging a rack 110 may be used to move the arm 36 along a rail arrangement 60. Preferably, the mobile element 36 will provide a precise positioning to the measuring unit 4. For example, a submersible rotary optical encoder 114 provided with a pinion 115 engaging the rack 110 may be used to measure the horizontal displacement of the mobile element 36. A proportional control unit may then control the position of the mobile element 36. In order to correct the position of the measuring unit 4 due to the inclination of the frame 6, a submersible inclinometer 118 measuring the inclination according to two degrees of liberty with respect to the horizontal may be used.
Referring to FIG. 4, the lasers 14, 16 of the measuring unit 4 are rigidly secured to the camera 22. Mirrors 46, 48 may be used to redirect the light planes 18, 20 toward the surface 2 depending on the chosen setup. Furthermore, the lasers 14, 16 may be disposed on a same side of the camera 22. Such an arrangement, shown in FIG. 5, allows reducing the size of the measuring unit 4. A single laser may also be used instead of two. In such a case, an appropriate optical arrangement (not shown) may be used to produce the two light planes 14, 16. Or, the light planes 14, 16 may be produced one after the other, so as to appear in subsequent images captured by the camera 22 and then processed in order to take this factor into account. It is also possible to use more than two laser projectors. In the case where n laser projectors are used, the intersection of the n light planes with the wires 10, 12 produces 2n points Pmi. Likewise, a laser projector may project several light planes on the surface 2 lighting as many points Pmi. FIG. 5 shows an arrangement of the measuring unit 4 where the laser projectors 14, 16 respectively project the light planes 17, 19 and 18, 20, lines 23, 25 and 24, 26 then resulting from the intersection of the light planes 17, 19 and 18, 20 with the surface 2.
The wires 10, 12 may be formed of steel wires or other types of wires if desired.
Referring to FIG. 5, it is also possible to use more than two wires, for example three wires 10, 11, 12 disposed in front of the surface 2, so that the light planes 24, 26 cross them. The extra wire or wires may be in the same plane as the wires 10, 12 and parallel to the wires 10, 12, or they may be in another plane, preferably parallel to the plane formed by the wires 10, 12, and still parallel to the wires 10, 12. The addition of extra wires provides more information for the position tracking of the measuring unit 4 and the on-line calibration of its parameters.
In order for the measuring unit 4 to self-track and map the surface 2, at least four points PM1, PM2, PM3, PM4 on the wires 10, 12 are lit by the laser arrangement 14, 16. An extra point on the wire 11 located at a distance g, which is not necessarily known, from the reference surface formed by the wires 10, 12 allows on-line calibrating the set of parameters of each light planes 18, 20.
Referring back to FIG. 1, the processing unit 8 may be embodied so as to build up a depth map of the surface 2 by means of an iterative and feedback process. For this purpose, the processing unit 8 comprises a tracking module 81, a mapping module 82 and a decision module 83. The tracking module 81 determines the location of the measuring unit 4 with respect to a fixed global reference system. The mapping module 82 determines the 3D coordinates of points on the surface 2 in the reference system bound to the camera 22 (shown in FIG. 3) of the measuring unit 4. The decision module 83 determines the commands to transmit to the support 6 that moves the measuring unit 4 in order to efficiently collect the data.
The tracking module 81 comprises two modeling sub-modules 56, 58, an estimation sub-module 812, a PMi points detection sub-module 810 and a distortions correction sub-module 811. The modeling sub-module 56 models the support assembly 6 by means of a set of parametric data that characterize it. Likewise, the modeling sub-module 58 models the measuring unit 4 by means of a set of parametric data that characterize it. The PMi points detection sub-module 810 determines the image coordinates of the points PM1, PM2, PM3, PM4. The distortions correction sub-module 811 corrects the image coordinates of the detected points in order to cancel the effect of the distortions of the lens 21 of the camera 22 or to take into account an out-of-parallelism defect of the wires 10, 12 (shown in FIG. 3). The estimation sub-module 812 estimates the pose (i.e. the position and orientation) of the measuring unit 4 with respect to the global reference system based on the signals provided by the modeling sub-module 58 and the distortions correction sub-module 811, and corrects the internal parameters of the measuring unit 4 through a feedback loop. The estimation sub-module 812 may use a Kalman filter to perform the desired estimation.
Referring to FIG. 3, the parametric data of the measuring unit 4 comprise the internal and external parameters of the camera 22, the parameters of the light planes 18, 20 and the spacing of the wires 10, 12. In the case where a perspective camera model with distortions is used, the typical internal parameters of the camera are then the focal lengths αuv, the main point c0=[u0,v0]T and the distortion coefficients of the lens 21. The typical external parameters of the camera 22 are the orientation and the position with respect to the reference system bound to the wires 10, 12. The light planes 18, 20 are defined by the parameters a, b, c, d in the equation in Cartesian coordinates: aX+bY+cZ+d=0. The parameter that defines the spacing of the wires 10, 12 is marked 2 h. Finally, the parametric data of the support assembly 6 (shown in FIG. 2) typically comprise the position and the orientation with respect to a fixed global reference system as well as the position of the measuring unit 4 on the transverse rail arrangement 60 (shown in FIG. 9).
Referring back to FIG. 1, the mapping module 82 comprises a Pu points detection sub-module 820 that determines the image coordinates of the points contained in the lines 24, 26, a distortions correction sub-module 811 similar to that of the tracking module 81, and a triangulation sub-module 822. The function of the triangulation sub-module 822 is to compute the 3D coordinates of the points Pu in the reference system bound to the camera 22 of the measuring unit 4. The mapping module 82 also comprises a reference change sub-module 840 that transforms the coordinates of the points Pu from the reference system bound to the camera toward the global reference system, and a surface modeling sub-module 54 that progressively builds up the map of the surface 2. There are numerous ways to represent the structural surface 2. For example, the surface modeling sub-module 54 may contain a set of 3D points that form the map of the surface 2 and define its tridimensional cartography.
The decision module 83 comprises a decision taking sub-module 830 that transmits positioning commands to the support assembly 6 and a communication interface 832 allowing initializing the processing unit 8 and retrieving the information sought, namely the map of the surface 2, the position and orientation data of the support assembly 6, and the parametric data of the measuring unit 4.
Through the wires 10, 12 (shown in FIG. 3), the measuring unit 4 of the system may be located with a great precision. The system also maps at the same time at least one line of points of the surface 2 by the use of a single image. For that, the camera 22 transmits the images 28 to the processing unit 8. The processing unit 8 processes the images 28 so as to locate the measuring unit 4 with respect to the reference system bound to the wires 10, 12 and so as to determine the coordinates of 3D points of the surface 2. The image coordinates of the points PM1, PM2, PM3, and PM4 are used to locate the measuring unit 4 while the lines 24, 26 are used to map the surface 2. The system according to the invention thus allows mapping surfaces of great length while preserving a high precision.
Referring also to FIGS. 2 and 3, the decision sub-module 83 initially transmits positioning commands to the support assembly 6 in order to position the measuring unit 4 at the level of the surface 2 to be inspected. When the measuring unit 4 is at the desired height, the decision sub-module 83 locks the displacement of the support assembly 6. The estimation sub-module 812 (e.g. Kalman filter) then corrects the internal operating parameters of the measuring unit 4 such as the parameters of the laser planes 18, 20. Then, the decision sub-module 83 transmits positioning commands to the support assembly 6 to move the measuring unit 4 crosswise (in the directions of arrows 42, 44 shown in FIG. 2) along the surface 2 in order to measure the 3D coordinates of the points contained in the lines 24, 26 of the surface 2. The mapping module 82 takes measurements at specified intervals via the communication interface 832. For each pair of lines 24, 26, an image is transmitted by the camera 22 to the mapping module 82. The surface modeling sub-module 54 builds up the map of the surface 2 line by line. The obtained map covers a strip of the surface 2 having a certain height. If several strips are required to map the whole or a desired part of the surface 2, then the decision module 83 transmits positioning commands to the support assembly 6 in order to move the measuring unit 4 heightwise toward the next adjoining area and then the above process is repeated.
The displacement of the measuring unit 4 may be made in a continuous mode or step by step. In the case of a continuous displacement, a single measurement for a given point of the surface 2 is taken by the tracking module 81 and by the mapping module 82. In the case of a step by step displacement, several measurements may be taken at each stop and may be combined so as to increase the confidence in the measurements.
In operation, the mapping system is likely to be subjected to multiple perturbations. Although it might be possible to directly take measurements of the structural surface 2 and then to reference these measurements using the position data of the measuring unit 4 and of the support assembly 6, these measurements are imprecise, or even false, due to the perturbations met. The system takes advantage of the wires 10, 12 through an iterative and feedback process, implemented in the tracking module 81, which ensures the validity of the position data of the measuring unit 4. This iterative and feedback process uses the model of the measuring unit 4 defined in the sub-module 58. Its role is to predict which will be the position of the points PM1, PM2, PM3, and PM4 in the image 28 based on the position and the orientation of the measuring unit 4, the internal parameters of the measuring unit 4 and the spacing of the wires 10, 12.
The minimal set of data for computing the pose of a camera without ambiguity is with four coplanar points. The PM1 points detector 810 determines the image coordinates of the points PM1, PM2, PM3, and PM4. The determination of the coordinates of the points PM1, PM2, PM3, and PM4 may be achieved using computer vision library algorithms. The distortions corrector 811 corrects these coordinates to cancel the effect of the possible distortions of the lens 21 of the camera 22. The points PM1, PM2, PM3, and PM4 are then triangulated in the reference system bound to the camera 22 in order to obtain their 3D coordinates. Equation 1 may be used for this purpose:
[ X Y Z ] CAM = [ ( u - u 0 ) Z / α u ( v - v 0 ) Z / α v - d / ( a + b ( u - u 0 ) / α u + c ( v - v 0 ) / α v ) ] CAM ( 1 )
where (u, v) represent the image coordinates of a point P, and (X, Y, Z) represent the 3D coordinates of P in the reference system bound to the camera 22. The orientation of the camera may be computed using the algorithm proposed by B. K. P. Horn in the paper “Closed-form solution of absolute orientation using unit quaternions”, Journal of the Optical Society of America A, Vol. 4, pages 629-642, 1987, with the 3D coordinates of the points PMI, PM2, PM3, and PM4 as input data. But other methods working on a set of 4 coplanar points may also be used. Once the orientation of the camera has been computed, the translation may be computed following the equation 2:
Trans(FILS,CAM)=−Rot(FILS,CAM)[ P Mi]CAM  (2)
where Trans(FILS, CAM) denotes the translation of the camera 22 with respect to the wires 10, 12, Rot(FILS, CAM) denotes the orientation of the camera 22 with respect to the wires 10, 12 and [ P Mi]CAM denotes the baric center of the points PM1, PM2, PM3, and PM4 expressed in the reference system of the camera 22. The position/orientation of the camera 22 thus computed comes as initial parameter of the Kalman filter 812 that refines the estimate of this position/orientation based on the parametric data of the model of the measuring unit 4 provided by the sub-module 58, the model of the support assembly 6 provided by the sub-module 56, and the computed pose. After several iterations of the Kalman filter 812, if the difference between the coordinates of the points PM1, PM2, PM3, et PM4 estimated by the model of the measuring unit 4 and the coordinates of the points PM1, PM2, PM3, and PM4 detected by the PMi points detector 810 is above a certain threshold, then the pose is rejected. Otherwise, the computed pose is valid. The use of this technique allows having a more robust and uncorrupted system. It is made possible by the presence of the wires 10, 12. This technique thus allows detecting if the wires 10, 12 are moving in response to perturbations or also if the parametric data of the model of the measuring unit 4 are false.
In the case where re-estimation of the parameters a, b, c, d of the laser planes 18, 20 is also desired, the third wire 11 (shown in FIG. 5) is used. These parameters are estimated by the Kalman filter 812 which has the position and the orientation of the camera 22 and the parameters a, b, c, d of the laser planes 18, 20 as state vector. The PMi points detector 810 determines the image coordinates of the points PM1, PM2, PM3, PM4, PM5, PM6 that are transmitted to the distortions corrector 811 and then to the Kalman filter 812. The position of the third wire 11 with respect to the wires 10, 12 is not necessarily known and may also be estimated by the Kalman filter 812.
The PLi points detector 820 determines the 2D coordinates of points of the images 28 contained in the lines 24, 26. This processing step may be achieved using computer vision library algorithms. The distortions corrector 811 then corrects the coordinates of the points PLi so as to cancel the effect of possible distortions of the lens 21 of the camera 22. In the case of a continuous displacement, the 3D coordinates of the points PLi, expressed in the reference system bound to the measuring unit 4, are determined from a single measurement by the triangulation sub-module 822 which uses the equation 1. In the case of a step by step displacement, it is possible, for example, to work out the average of the points PLi measured several times by the sub-modules 820, 811 and 822. A change of reference system 840 is then achieved to obtain the coordinates of Pu in the global reference system. The reference change may be achieved following the equation:
[ X Y Z 1 ] GLOBAL = T ( GLOBAL , CAM ) [ X Y Z 1 ] CAM ( 3 )
where T is a homogenous transformation matrix that is function of the orientation and of the position of the camera 22 with respect to the global reference system.
The positioning of the measuring unit 4 on the support assembly 6 may be performed by the decision taking sub-module 830 in order to be able to control the positioning as function of the processed measurements. For example, it may be possible that the system encounters perturbations (for example, motions of the wires 10, 12) and is no longer able to take measurements. The decision taking sub-module 830 may in that case stop the displacement of the measuring unit 4 until the perturbations disappear. The positioning may also be pre-programmed so as to be achieved without feedback. Or an operator may take control of the displacements via the communication interface 832.
The lines 24, 26 may be indexed by means of an optical encoder 114 (shown in FIG. 9) that provides the position of the measuring unit 4 on the transverse rail arrangement 60 of the support assembly 6 (shown in FIG. 2). The index of the lines 24, 26 corresponds to the position of the measuring unit 4 on the rail arrangement 60 of the support assembly 6.
The configuration of the tracking, mapping and decision modules 81, 82, 83 may be adapted for example to refine certain processings in order to produce the mapping data or according to the operating parameters and the type of equipment forming the measuring unit 4 and the support assembly 6.
Referring to FIG. 7, the processing unit 8 and the communication interface 832 (shown in FIG. 1) may advantageously be implemented in a processing device 62 such as a PC (computer) or appropriate dedicated circuits controlling the functions of the system. The processing device 62 is connected to the measuring unit 4 and retrieves the images 28 from the camera 22 (shown in FIG. 2) and processes them to locate the measuring unit 4 and map the surface 2. The processing may be achieved in continuous mode at the refresh rate of the camera 22. The processing device 62 may also control the two displacement axes of the support assembly 6. During a translation of the measuring unit 4 along the horizontal axis ( arrows 42, 44 in FIG. 3) and at specified intervals, the processing device 62 retrieves the position (X, Y) of the measuring unit 4 and stores a profile line in memory. This information is used to express the coordinates of the laser lines in a global reference system bound to the scene.
Referring back to FIG. 3, in order to be able to detect surface defects of the inspected structure of reduced sizes (0.1 mm for example), the camera 22 is preferably provided with a high resolution progressive CCD matrix. Such a camera may communicate with the processing device 62 (shown in FIG. 7) via a digital transmission channel such as IEEE-1394 (also known under the denomination FireWire), Camera Link, USB or Gigabyte Ethernet. The laser projectors 14, 16 preferably emit at a wavelength in the visible range (between 400 nm and 800 nm). To facilitate the detection of the lines 24, 26 in the image 28, the lens 21 of the camera 22 may use an optical filter that let only the wavelength of the laser projectors 14, 16 through.
Although the system is designed to operate underwater, it may also be used outside of water if desired. In such a case, the calibration of the measuring unit 4 is simply made outside of water, and the system then works in the same way as underwater.
While embodiments of the invention have been illustrated in the accompanying drawings and described above, it will be evident to those skilled in the art that modifications may be made therein without departing from the invention.

Claims (16)

The invention claimed is:
1. A system for tridimensional cartography of a structural surface, comprising:
two wires extendable along the structural surface while being spaced from each other so as to define a reference surface in front of the structural surface;
a measuring unit comprising a laser arrangement configured to project two distinct light planes directed toward the structural surface and intersecting with the wires at an intersection, and a camera configured to capture images of the structural surface containing lines resulting from an intersection of the light planes with the structural surface and reference points resulting from the intersection of the light planes with the wires;
a support assembly configured to controllably move the measuring unit in front of the structural surface so as to progressively scan the structural surface to be mapped; and
a processing unit for processing the images captured by the camera, the processing unit having:
a processing circuit that determines an orientation and a position of the measuring unit with respect to the reference surface as function of the reference points in the images;
a processing circuit that determines 3D coordinates of points defining the lines in the images with respect to a reference system bound to the measuring unit determined as function of the orientation and the position of the measuring unit; and
a processing circuit that transforms the 3D coordinates into cartographic data representative of the structural surface in a global reference system bound to the reference surface defined by the wires.
2. The system according to claim 1, wherein the support assembly comprises a frame having a mobile element to which the measuring unit is attached, and an arrangement of transverse and vertical guiding elements for the mobile element.
3. The system according to claim 2, wherein the arrangement of guiding elements comprises lateral guides spaced from each other and configured to extend in front of the structural surface, a transverse element slideably mounted on the lateral guides and slideably receiving the mobile element, and an arrangement for displacement of the mobile element with respect to the transverse element and of the transverse element with respect to the lateral guides.
4. The system according to claim 3, wherein the wires extend between the lateral guides.
5. The system according to claim 1, wherein the laser arrangement comprises two lasers attached to the camera.
6. The system according to claim 1, further comprising an extra wire extendable along the structural surface so that the light planes cross the extra wire, and wherein the processing unit further has a processing circuit that calibrates parameters of the light planes as function of a reference point in the images captured by the camera corresponding to an intersection of one of the light planes with the extra wire.
7. The system according to claim 1, wherein:
the processing circuit that determines the orientation and the position comprises a tracking module having:
a support assembly modeling sub-module that models the support assembly through a set of parametric data characterizing the support assembly;
a measuring unit modeling sub-module that models the measuring unit through a set of parametric data characterizing the measuring unit;
a sub-module for detection of the reference points in the images, that determines the image coordinates of the reference points;
a distortions correction sub-module that corrects the image coordinates of the reference points based on the parametric data characterizing a distortion relative to the measuring unit; and
an estimation sub-module that estimates the position and the orientation of the measuring unit in the global reference system based on the parametric data characterizing the measuring unit and the corrected image coordinates, and corrects the parametric data characterizing the measuring unit based on the estimated position and the orientation of the measuring unit; and
the processing circuit that determines the 3D coordinates of the points and the processing circuit that transforms the 3D coordinates comprise together a mapping module having:
a sub-module for detection of points defining the lines in the images, that determines image coordinates of the points defining the lines;
a distortions correction sub-module that corrects the image coordinates of the points defining the lines based on the parametric data characterizing the distortion relative to the measuring unit;
a triangulation sub-module that determines the 3D coordinates of the points defining the lines from the reference system bound to the measuring unit;
a reference change sub-module that transforms the 3D coordinates of the points defining the lines from the reference system bound to the measuring unit toward the global reference system to produce the mapping data; and
a structural surface modeling module that builds up a map of the surface based on to the mapping data.
8. The system according to claim 7, wherein the processing unit further comprises a decision module having:
a decision taking sub-module connected to the support assembly modeling sub-module and to the support assembly, the decision taking sub-module transmitting positioning commands to the support assembly based on the mapping data in the surface map; and
a communication interface configured to receive operating data directed to the processing unit and to transmit the map of the surface, data representative of the position and orientation of the measuring unit and the parametric data characterizing the measuring unit.
9. A method for tridimensional cartography of a structural surface, comprising the steps of:
a) extending two wires along the structural surface while spacing them from each other so as to define a reference surface in front of the structural surface;
b) controllably moving a measuring unit comprising a laser arrangement and a camera in front of the structural surface so as to progressively scan the structural surface;
c) projecting two distinct light planes directed toward the structural surface using the laser arrangement;
d) capturing images of the structural surface containing lines resulting from an intersection of the light planes with the structural surface and four reference points resulting from an intersection of the light planes with the wire using the camera;
e) determining an orientation and a position of the measuring unit with respect to the reference surface as function of the reference points in the images;
f) determining 3D coordinates of points defining the lines in the images with respect to a reference system bound to the measuring unit determined as function of the orientation and the position of the measuring unit; and
g) transforming the 3D coordinates into cartographic data representative of the structural surface in a reference system bound to the reference surface defined by the wires.
10. The method according to claim 9, wherein the step of moving comprises a transmission of positioning commands directed to move the measuring unit based on the mapping data gathered.
11. The method according to claim 9, wherein the laser arrangement is rigidly secured to the camera.
12. The method according to claim 9, further comprising the steps of:
extending an extra wire along the structural surface so that the light planes cross the extra wire; and
calibrating parameters of the light planes as function of a reference point in the images captured by the camera corresponding to an intersection of one of the light planes with the extra wire.
13. The method according to claim 9, wherein the wires extend close to the structural surface and in parallel with each other, the reference surface being planar.
14. The method according to claim 9, wherein the step of determining the orientation and the position of the measuring unit comprises the sub-steps of:
modeling a support assembly configured to controllably move the measuring unit in front of the structural surface by a set of parametric data characterizing the support assembly;
modeling the measuring unit by a set of parametric data characterizing the measuring unit;
detecting the reference points in the images and determining image coordinates of the reference points;
correcting the image coordinates of the reference points based on the parametric data characterizing a distortion relative to the measuring unit;
estimating the position and the orientation of the measuring unit in the global reference system based on the parametric data characterizing the measuring unit and the corrected image coordinates; and
correcting the parametric data characterizing the measuring unit based on the estimated position and orientation of the measuring unit.
15. The method according to claim 9, wherein the step of determining the 3D coordinates of the points defining the lines in the images comprises the sub-steps of:
detecting the points defining the lines in the images and determining image coordinates of the points defining the lines;
correcting the image coordinates of the points defining the lines based on the parametric data characterizing a distortion relative to the measuring unit; and
determining by triangulation the 3D coordinates of the points defining the lines in the reference system bound to the measuring unit.
16. The method according to claim 9, wherein the step of transforming comprises the sub-steps of:
producing the mapping data by transformation of the 3D coordinates of the points defining the lines from the reference system bound to the measuring unit toward the global reference system; and
building up a map of the surface based on the mapping data.
US12/734,034 2007-10-11 2008-10-08 System and method for tridimensional cartography of a structural surface Active 2030-06-30 US8462208B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CA002606267A CA2606267A1 (en) 2007-10-11 2007-10-11 System and method for three-dimensional mapping of a structural surface
CA2,606,267 2007-10-11
PCT/CA2008/001769 WO2009046519A1 (en) 2007-10-11 2008-10-08 System and method for the three-dimensional mapping of a structural surface

Publications (2)

Publication Number Publication Date
US20100238269A1 US20100238269A1 (en) 2010-09-23
US8462208B2 true US8462208B2 (en) 2013-06-11

Family

ID=40527649

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/734,034 Active 2030-06-30 US8462208B2 (en) 2007-10-11 2008-10-08 System and method for tridimensional cartography of a structural surface

Country Status (3)

Country Link
US (1) US8462208B2 (en)
CA (2) CA2606267A1 (en)
WO (1) WO2009046519A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11506484B2 (en) * 2011-11-30 2022-11-22 Nikon Corporation Profile measuring apparatus, structure manufacturing system, method for measuring profile, method for manufacturing structure, and non-transitory computer readable medium

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AT507618B1 (en) * 2008-11-26 2012-01-15 Riegl Laser Measurement Sys METHOD FOR DETERMINING THE RELATIVE POSITION OF A LASER SCANNER TO A REFERENCE SYSTEM
CA2694883C (en) 2010-02-25 2016-08-09 Hydro-Quebec Robot for tooling a structural part under water
BR112015013804B1 (en) * 2012-12-14 2021-02-09 Bp Corporation North America Inc measuring system for three-dimensional measurement of an underwater structure, method for laser triangulation of an underwater structure and non-transient computer-readable medium coded with instructions
CA2908738A1 (en) * 2013-04-05 2014-10-09 Lockheed Martin Corporation Underwater platform with lidar and related methods
MX2013010928A (en) * 2013-09-24 2015-03-23 Generadora De Desarrollos Tecnologicos Gdtec S De R L De C V Three-dimensional image scanner.
PL233457B1 (en) * 2017-10-01 2019-10-31 Politechnika Rzeszowska Im Ignacego Lukasiewicza Method for determining value of the angle of inclination of dendrite cores in relation of the surface of a mould in monocrystalline castings
CN109778789B (en) * 2018-04-04 2020-08-25 水利部交通运输部国家能源局南京水利科学研究院 Concrete panel rock-fill dam panel goaf underwater repair system
FR3118487B1 (en) * 2020-12-29 2023-02-10 Naval Group SENSOR STRUCTURE
CN113109762B (en) * 2021-04-07 2022-08-02 哈尔滨工程大学 Optical vision guiding method for AUV (autonomous Underwater vehicle) docking recovery
CN115452820B (en) * 2022-07-21 2023-10-27 成都华建地质工程科技有限公司 Method, device and medium for extracting structural surface characteristics based on borehole television image

Citations (48)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4175862A (en) 1975-08-27 1979-11-27 Solid Photography Inc. Arrangement for sensing the geometric characteristics of an object
US4294544A (en) 1979-08-03 1981-10-13 Altschuler Bruce R Topographic comparator
US4492471A (en) * 1980-04-23 1985-01-08 Pharos Ab Arrangement for checking dimensional accuracy
US4612797A (en) 1984-06-27 1986-09-23 Rockwell International Corporation Leak locating and mapping system and method
US4811250A (en) * 1986-05-02 1989-03-07 Applied Power Inc. Deviation measurement system
US4981353A (en) 1988-05-06 1991-01-01 Mitsubishi Jukogyo Kabushiki Kaisha Position locating apparatus for an underwater moving body
US5162861A (en) * 1991-04-01 1992-11-10 Tamburino Louis A Laser imaging and ranging system using one camera
US5679899A (en) 1995-03-06 1997-10-21 Holographics Inc. Method and apparatus for non-destructive testing of structures
US5898181A (en) * 1995-06-30 1999-04-27 Hdi Instrumentation Thin film optical measurement system
US6122541A (en) * 1995-05-04 2000-09-19 Radionics, Inc. Head band for frameless stereotactic registration
US6268923B1 (en) * 1999-10-07 2001-07-31 Integral Vision, Inc. Optical method and system for measuring three-dimensional surface topography of an object having a surface contour
US6379043B1 (en) * 1998-12-08 2002-04-30 U.S. Philips Corporation X-ray examination apparatus and method for generating distortion-free X-ray images
US6510244B2 (en) * 1996-06-13 2003-01-21 K. U. Leuven Research & Development Method and system for acquiring a three-dimensional shape description
US6549288B1 (en) * 1998-05-14 2003-04-15 Viewpoint Corp. Structured-light, triangulation-based three-dimensional digitizer
US6574050B1 (en) * 2001-12-07 2003-06-03 Industrial Technology Research Institute 3D shape-measuring apparatus using biaxial anamorphic magnification
US6577405B2 (en) * 2000-01-07 2003-06-10 Cyberoptics Corporation Phase profilometry system with telecentric projector
US6608913B1 (en) * 2000-07-17 2003-08-19 Inco Limited Self-contained mapping and positioning system utilizing point cloud data
US6610992B1 (en) * 2000-07-19 2003-08-26 Clasmet Rotating beam method and system for measuring part edges and openings
US20030164952A1 (en) * 2000-08-25 2003-09-04 Nikolaj Deichmann Method and apparatus for three-dimensional optical scanning of interior surfaces
US6677941B2 (en) * 2000-08-05 2004-01-13 American Gnc Corporation Three-dimensional relative positioning and tracking using LDRI
WO2004011876A1 (en) 2002-07-25 2004-02-05 Solutionix Corporation Apparatus and method for automatically arranging three dimensional scan data using optical marker
US6703634B2 (en) * 2000-12-11 2004-03-09 Matsushita Electric Industrial Co., Ltd. 3D-shape measurement apparatus
US6714307B2 (en) * 2001-10-16 2004-03-30 Zygo Corporation Measurement of complex surface shapes using a spherical wavefront
US6823737B2 (en) * 2001-06-06 2004-11-30 The United States Of America As Represented By The Secretary Of The Interior Non-contact inspection system for large concrete structures
US6885479B1 (en) * 1999-07-09 2005-04-26 Hewlett-Packard Development Company, L.P. Curled surface imaging system
EP1062478B1 (en) 1998-03-13 2005-06-01 Metroptic Technologies, Ltd. Apparatus and method for optically measuring an object surface contour
US20060018642A1 (en) 2004-07-20 2006-01-26 Chaplin David V Mobile laser designated infrared multimedia mapping system
US20060093206A1 (en) 2000-04-28 2006-05-04 Rudger Rubbert System and method for mapping a surface
US7075661B2 (en) * 2001-02-23 2006-07-11 Industrial Control Systems Limited Apparatus and method for obtaining three-dimensional positional data from a two-dimensional captured image
US20070115484A1 (en) * 2005-10-24 2007-05-24 Peisen Huang 3d shape measurement system and method including fast three-step phase shifting, error compensation and calibration
US7248344B2 (en) * 2001-04-04 2007-07-24 Instro Precision Limited Surface profile measurement
US7253908B2 (en) 2004-07-22 2007-08-07 The Boeing Company Non-destructive inspection using laser profiling and associated method
US7259871B2 (en) 2003-06-12 2007-08-21 National Taipei University Of Technology Apparatus and method for rapid and precise scanning of three-dimensional occlusal profile of dental cast
US20070206204A1 (en) 2005-12-01 2007-09-06 Peirong Jia Full-field three-dimensional measurement method
US7336375B1 (en) * 2006-10-04 2008-02-26 Ivan Faul Wireless methods and systems for three-dimensional non-contact shape sensing
US20080055588A1 (en) 2006-09-01 2008-03-06 Disco Corporation Via hole depth detector
US7345757B2 (en) 2004-09-28 2008-03-18 Penza G Gregory Inspection apparatus for pipelines
US7352455B2 (en) * 2002-01-09 2008-04-01 Chief Automotive Technologies, Inc. Laser scanner with parabolic collector
US20080088848A1 (en) 2006-10-16 2008-04-17 Hitachi High-Technologies Corporation Foreign matter inspection apparatus and foreign matter inspection method
US7392151B2 (en) * 2003-03-25 2008-06-24 Sandvik Mining And Construction Oy Initializing position and direction of mining vehicle
US7489408B2 (en) * 2005-11-15 2009-02-10 General Electric Company Optical edge break gage
US7609388B2 (en) * 2002-01-24 2009-10-27 Icos Vision Systems Nv Spatial wavefront analysis and 3D measurement
US7747312B2 (en) * 2000-01-04 2010-06-29 George Mason Intellectual Properties, Inc. System and method for automatic shape registration and instrument tracking
US7768656B2 (en) * 2007-08-28 2010-08-03 Artec Group, Inc. System and method for three-dimensional measurement of the shape of material objects
US7899221B2 (en) * 2001-11-08 2011-03-01 Institut Straumann Ag Devices and methods for producing denture parts
US7912673B2 (en) * 2005-03-11 2011-03-22 Creaform Inc. Auto-referenced system and apparatus for three-dimensional scanning
US7995834B1 (en) * 2006-01-20 2011-08-09 Nextengine, Inc. Multiple laser scanner
US8014002B2 (en) * 2008-04-01 2011-09-06 Perceptron, Inc. Contour sensor incorporating MEMS mirrors

Patent Citations (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4175862A (en) 1975-08-27 1979-11-27 Solid Photography Inc. Arrangement for sensing the geometric characteristics of an object
US4294544A (en) 1979-08-03 1981-10-13 Altschuler Bruce R Topographic comparator
US4492471A (en) * 1980-04-23 1985-01-08 Pharos Ab Arrangement for checking dimensional accuracy
US4612797A (en) 1984-06-27 1986-09-23 Rockwell International Corporation Leak locating and mapping system and method
US4811250A (en) * 1986-05-02 1989-03-07 Applied Power Inc. Deviation measurement system
US4981353A (en) 1988-05-06 1991-01-01 Mitsubishi Jukogyo Kabushiki Kaisha Position locating apparatus for an underwater moving body
US5162861A (en) * 1991-04-01 1992-11-10 Tamburino Louis A Laser imaging and ranging system using one camera
US5679899A (en) 1995-03-06 1997-10-21 Holographics Inc. Method and apparatus for non-destructive testing of structures
US6122541A (en) * 1995-05-04 2000-09-19 Radionics, Inc. Head band for frameless stereotactic registration
US5898181A (en) * 1995-06-30 1999-04-27 Hdi Instrumentation Thin film optical measurement system
US6510244B2 (en) * 1996-06-13 2003-01-21 K. U. Leuven Research & Development Method and system for acquiring a three-dimensional shape description
EP1062478B1 (en) 1998-03-13 2005-06-01 Metroptic Technologies, Ltd. Apparatus and method for optically measuring an object surface contour
US6549288B1 (en) * 1998-05-14 2003-04-15 Viewpoint Corp. Structured-light, triangulation-based three-dimensional digitizer
US6379043B1 (en) * 1998-12-08 2002-04-30 U.S. Philips Corporation X-ray examination apparatus and method for generating distortion-free X-ray images
US6885479B1 (en) * 1999-07-09 2005-04-26 Hewlett-Packard Development Company, L.P. Curled surface imaging system
US6268923B1 (en) * 1999-10-07 2001-07-31 Integral Vision, Inc. Optical method and system for measuring three-dimensional surface topography of an object having a surface contour
US7747312B2 (en) * 2000-01-04 2010-06-29 George Mason Intellectual Properties, Inc. System and method for automatic shape registration and instrument tracking
US6577405B2 (en) * 2000-01-07 2003-06-10 Cyberoptics Corporation Phase profilometry system with telecentric projector
US20060093206A1 (en) 2000-04-28 2006-05-04 Rudger Rubbert System and method for mapping a surface
US6608913B1 (en) * 2000-07-17 2003-08-19 Inco Limited Self-contained mapping and positioning system utilizing point cloud data
US6610992B1 (en) * 2000-07-19 2003-08-26 Clasmet Rotating beam method and system for measuring part edges and openings
US6677941B2 (en) * 2000-08-05 2004-01-13 American Gnc Corporation Three-dimensional relative positioning and tracking using LDRI
US20030164952A1 (en) * 2000-08-25 2003-09-04 Nikolaj Deichmann Method and apparatus for three-dimensional optical scanning of interior surfaces
US6703634B2 (en) * 2000-12-11 2004-03-09 Matsushita Electric Industrial Co., Ltd. 3D-shape measurement apparatus
US7075661B2 (en) * 2001-02-23 2006-07-11 Industrial Control Systems Limited Apparatus and method for obtaining three-dimensional positional data from a two-dimensional captured image
US7248344B2 (en) * 2001-04-04 2007-07-24 Instro Precision Limited Surface profile measurement
US6823737B2 (en) * 2001-06-06 2004-11-30 The United States Of America As Represented By The Secretary Of The Interior Non-contact inspection system for large concrete structures
US6714307B2 (en) * 2001-10-16 2004-03-30 Zygo Corporation Measurement of complex surface shapes using a spherical wavefront
US7899221B2 (en) * 2001-11-08 2011-03-01 Institut Straumann Ag Devices and methods for producing denture parts
US6574050B1 (en) * 2001-12-07 2003-06-03 Industrial Technology Research Institute 3D shape-measuring apparatus using biaxial anamorphic magnification
US7352455B2 (en) * 2002-01-09 2008-04-01 Chief Automotive Technologies, Inc. Laser scanner with parabolic collector
US7609388B2 (en) * 2002-01-24 2009-10-27 Icos Vision Systems Nv Spatial wavefront analysis and 3D measurement
WO2004011876A1 (en) 2002-07-25 2004-02-05 Solutionix Corporation Apparatus and method for automatically arranging three dimensional scan data using optical marker
US7392151B2 (en) * 2003-03-25 2008-06-24 Sandvik Mining And Construction Oy Initializing position and direction of mining vehicle
US7259871B2 (en) 2003-06-12 2007-08-21 National Taipei University Of Technology Apparatus and method for rapid and precise scanning of three-dimensional occlusal profile of dental cast
US20060018642A1 (en) 2004-07-20 2006-01-26 Chaplin David V Mobile laser designated infrared multimedia mapping system
US7253908B2 (en) 2004-07-22 2007-08-07 The Boeing Company Non-destructive inspection using laser profiling and associated method
US7345757B2 (en) 2004-09-28 2008-03-18 Penza G Gregory Inspection apparatus for pipelines
US7912673B2 (en) * 2005-03-11 2011-03-22 Creaform Inc. Auto-referenced system and apparatus for three-dimensional scanning
US20070115484A1 (en) * 2005-10-24 2007-05-24 Peisen Huang 3d shape measurement system and method including fast three-step phase shifting, error compensation and calibration
US7489408B2 (en) * 2005-11-15 2009-02-10 General Electric Company Optical edge break gage
US20070206204A1 (en) 2005-12-01 2007-09-06 Peirong Jia Full-field three-dimensional measurement method
US7995834B1 (en) * 2006-01-20 2011-08-09 Nextengine, Inc. Multiple laser scanner
US7471384B2 (en) * 2006-09-01 2008-12-30 Disco Corporation Via hole depth detector
US20080055588A1 (en) 2006-09-01 2008-03-06 Disco Corporation Via hole depth detector
US7336375B1 (en) * 2006-10-04 2008-02-26 Ivan Faul Wireless methods and systems for three-dimensional non-contact shape sensing
US20080088848A1 (en) 2006-10-16 2008-04-17 Hitachi High-Technologies Corporation Foreign matter inspection apparatus and foreign matter inspection method
US7768656B2 (en) * 2007-08-28 2010-08-03 Artec Group, Inc. System and method for three-dimensional measurement of the shape of material objects
US8014002B2 (en) * 2008-04-01 2011-09-06 Perceptron, Inc. Contour sensor incorporating MEMS mirrors

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11506484B2 (en) * 2011-11-30 2022-11-22 Nikon Corporation Profile measuring apparatus, structure manufacturing system, method for measuring profile, method for manufacturing structure, and non-transitory computer readable medium

Also Published As

Publication number Publication date
US20100238269A1 (en) 2010-09-23
CA2606267A1 (en) 2009-04-11
CA2701497A1 (en) 2009-04-16
WO2009046519A1 (en) 2009-04-16
CA2701497C (en) 2014-12-09

Similar Documents

Publication Publication Date Title
US8462208B2 (en) System and method for tridimensional cartography of a structural surface
WO2021135157A1 (en) Rock mass structure detection and dangerous rock detection system and method
US9223025B2 (en) Underwater platform with LIDAR and related methods
US10397550B2 (en) Apparatus and method for three dimensional surface measurement
US10739274B2 (en) Imaging support device and imaging support method
US20100226541A1 (en) System and method for detecting position of underwater vehicle
US10930013B2 (en) Method and system for calibrating imaging system
Cho et al. Target-focused local workspace modeling for construction automation applications
RU2581103C1 (en) System for control over assembly steps in floating dock
Menna et al. Towards real-time underwater photogrammetry for subsea metrology applications
US20220102018A1 (en) Systems and methods for laser inspection and measurements
Mirallès et al. Laser scanning system for inspecting large underwater hydroelectric structures
KR100516395B1 (en) Visual sensor system and environment modeling method for autonomous mobile welding robots
US20230073689A1 (en) Inspection Device for Inspecting a Building or Structure
JP4855547B1 (en) Method for analyzing the configuration of stereotyped bodies
Nunes et al. Occupancy grid mapping from 2d sonar data for underwater scenes
Menna et al. Evaluation of vision-based localization and mapping techniques in a subsea metrology scenario
CN110625614A (en) 2D laser positioning method for train body door and window
Roman et al. Estimation of error in large area underwater photomosaics using vehicle navigation data
Loeffler et al. The calibration of a laser profiling system for seafloor micro-topography measurements
CN117470186A (en) Laser collimation type sedimentation automatic monitoring method and device
CN117516545A (en) Heading machine navigation positioning method, system and device based on ultrasonic technology
Thielemann et al. Robot navigation and obstacle detection in pipelines using time-of-flight imagery
CN114705682A (en) Intelligent visual detection, identification and imaging device for rock mass structure
Mirallès et al. A 3D imaging system for inspection of large underwater hydroelectric structures

Legal Events

Date Code Title Description
AS Assignment

Owner name: HYDRO-QUEBEC, CANADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MIRALLES, FRANCOIS;HOUDE, REGIS;BEAUDRY, JULIEN;AND OTHERS;SIGNING DATES FROM 20100325 TO 20100330;REEL/FRAME:024204/0128

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8