US6556706B1 - Three-dimensional surface profile imaging method and apparatus using single spectral light condition - Google Patents

Three-dimensional surface profile imaging method and apparatus using single spectral light condition Download PDF

Info

Publication number
US6556706B1
US6556706B1 US09/770,124 US77012401A US6556706B1 US 6556706 B1 US6556706 B1 US 6556706B1 US 77012401 A US77012401 A US 77012401A US 6556706 B1 US6556706 B1 US 6556706B1
Authority
US
United States
Prior art keywords
light
camera
scene
wavelength
light source
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US09/770,124
Inventor
Z. Jason Geng
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Technest Holdings Inc
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US09/770,124 priority Critical patent/US6556706B1/en
Priority to PCT/US2001/018644 priority patent/WO2002059545A1/en
Priority to EP01946201A priority patent/EP1354176A1/en
Priority to AU2001268281A priority patent/AU2001268281A1/en
Publication of US6556706B1 publication Critical patent/US6556706B1/en
Application granted granted Critical
Assigned to GENEX TECHNOLOGIES, INC. reassignment GENEX TECHNOLOGIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GENG, ZHENG JASON
Assigned to SILICON VALLEY BANK reassignment SILICON VALLEY BANK SECURITY AGREEMENT Assignors: E-OIR TECHNOLOGIES, INC., GENEX TECHNOLOGIES INCORPORATED, TECHNEST HOLDINGS, INC.
Assigned to TECHNEST HOLDINGS, INC. reassignment TECHNEST HOLDINGS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GENEX TECHNOLOGIES, INC.
Assigned to E-OIR TECHNOLOGIES, INC., GENEX TECHNOLOGIES INCORPORATED, TECHNEST HOLDINGS, INC. reassignment E-OIR TECHNOLOGIES, INC. RELEASE Assignors: SILICON VALLEY BANK
Adjusted expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/24Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
    • G01B11/25Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
    • G01B11/2509Color coding

Definitions

  • the present invention is directed to three-dimensional surface profile imaging, and more particularly to a method and apparatus for three-dimensional imaging that uses color ranging to conduct surface profile measurement.
  • a three dimensional surface profile imaging method and apparatus described in U.S. Pat. No. 5,675,407 (“the '407 patent”), the disclosure of which is incorporated herein by reference in its entirety, conducts imaging by projecting light through a linear variable wavelength filter (LVWF), thereby projecting light having a known, spatially distributed wavelength spectrum on the objects being imaged.
  • the LVWF is a rectangular optical glass plate coated with a color-filtering film that gradually varies in color, (i.e., wavelength). If the color spectrum of a LVWF is within the visible light region, one edge of the filter rectangle may correspond to the shortest visible wavelength (i.e. blue or violet) while the opposite edge may correspond to the longest visible wavelength, (i.e. red).
  • the wavelength of light passing through the coated color-filtering layer is linearly proportional to the distance between the position on the filter glass where the light passes and the blue or red edge. Consequently, the color of the light is directly related to the angle ⁇ , shown in FIG. 1, at which the light leaves the rainbow projector and LVWF.
  • the imaging method and apparatus is based on the triangulation principle and the relationship between a light projector 100 that projects through the LVWF 101 , a camera 102 , and the object or scene being imaged 104 .
  • a triangle is uniquely defined by the angles theta ( ⁇ ) and alpha ( ⁇ ), and the length of the baseline (B).
  • the distance i.e., the range R
  • the key to the triangulation method is to determine the projection angle, ⁇ , from an image captured by the camera 102 and more particularly to determine all ⁇ angles corresponding to all the visible points on an object's surface in order to obtain a full-frame 3D image in one snapshot.
  • FIG. 2 is a more detailed version of FIG. 1 and illustrates the manner in which all visible points on the object's surface 104 is obtained via the triangulation method.
  • the light projector 100 generates a fan beam of light 200 .
  • the fan beam 200 is broad spectrum light (i.e., white light) which passes through the LVWF 101 to illuminate one or more three-dimensional objects 104 in the scene with a pattern of light rays possessing a rainbow-like spectrum distribution.
  • the fan beam of light 200 is composed of multiple vertical planes of light 202 , or “light sheets”, each plane having a given projection angle and wavelength.
  • the light reflected from the object 104 surface is then detected by the camera 102 .
  • a visible spectrum range LVWF 400-700 nm
  • the color detected by the camera pixels is determined by the proportion of its primary color Red, Green, and Blue components (RGB).
  • RGB Red, Green, and Blue components
  • the color spectrum of each pixel has a one-to-one correspondence with the projection angle ( ⁇ ) of the plane of light due to the fixed geometry of the camera 102 lens and the LVWF 101 characteristics. Therefore, the color of light received by the camera 102 can be used to determine the angle ⁇ at which that light left the light projector 100 through the LVWF 101 .
  • the angle ⁇ is determined by the physical relationship between the camera 102 and the coordinates of each pixel on the camera's imaging plane.
  • the baseline B between the camera's 102 focal point and the center of the cylindrical lens of the light projector 100 is fixed and known. Given the value for angles ⁇ and ⁇ , together with the known baseline length B, all necessary information is provided to easily determine the full frame of three-dimensional range values (x,y,z) for any and every visible spot on the surface of the objects 104 seen by the camera 102 .
  • the three-dimensional imaging system described above can capture full-frame, high spatial resolution three-dimensional images using a standard camera, such as a charge coupled device camera, in real time without relying on any moving parts.
  • a standard camera such as a charge coupled device camera
  • the imaging system does not rely on a laser, it does not pose any hazard to the eyes when used in clinical applications.
  • the wavelength of the light projected onto the object surface continuously varies, there is no theoretical limitation on the measurement accuracy that can be achieved by the system. The actual accuracy of a specific system will depend on system implementation and will be affected primarily by limiting factors such as the optical system design, the quality and resolution of the camera, light spectral emission of the light source projector; noise level and resolution of the frame grabber, calibration algorithms, and the three-dimensional imaging processing algorithms.
  • the system may obtain an image of the object under normal light conditions before projecting the filtered light onto the object.
  • the image obtained under normal light conditions is then subtracted from the image obtained under LVWF light conditions to eliminate the effects of the object color on the image.
  • the consistency of the spectral power distribution and the RGB value of each pixel may vary when light is projected onto the object through the LVWF based on the reflection characteristics of the object's surface, particularly if the object is not white and/or not uniformly colored.
  • the present invention is directed to a method and apparatus for three-dimensional surface imaging that avoids variations in the RGB value of each pixel due to the reflection characteristics of the object's surface. More particularly, a light source in the system illuminates an object or scene with a light pattern having a spatially varying wavelength and composed of at least one light plane. The light plane corresponds to at least one angle at which the light of that wavelength is emitted and contains only a single spectral component.
  • the RGB values of each pixel will be independent of the light intensity of the light source and the reflectance characteristics of the object or scene being imaged.
  • any color matching function that is conducted to link the wavelength of the light projected on the object or scene at a given point and that point's position will be consistent, regardless of the color and/or reflectance characteristics of the object's or scene's surface.
  • FIG. 1 is a simplified diagram illustrating a triangulation principle used in the present invention
  • FIG. 2 is a representative diagram of the components used by the inventive system.
  • FIG. 3 is a plan view of the system shown in FIG. 2 .
  • the inventive system and method system ensures that the RGB values corresponding to a given spectral composition f( ⁇ ) are consistent, regardless of the reflectance characteristics of the surface being imaged.
  • the invention recognizes that the spectral composition f( ⁇ ) of the light source 100 is a parameter that is controllable in the system shown in FIG. 2 . Because this parameter is controllable, imposing a restrictive condition on f( ⁇ ) can make the RGB values consistent for f( ⁇ ), regardless of r( ⁇ ) and s( ⁇ ). This will be explained in greater detail below.
  • the pixel values are computed from the integral of the spectral power distribution (SPD) of the light measured by a charge coupled device (CCD) element in the camera 102 .
  • SPD spectral power distribution
  • CCD charge coupled device
  • the camera 102 receives light directly from a light source
  • f( ⁇ ) is the spectral composition (spectral power distribution) of illumination source
  • s( ⁇ ) is the spectral sensitivity function of the sensor
  • ⁇ max and ⁇ min are the upper and lower bounds of the wavelengths visible to CCD pixels in the CCD elements.
  • r( ⁇ ) is the surface spectral reflectance function of the object in the scene 104 .
  • each filter has a transmittance function ⁇ ( ⁇ ), specifying the fraction of light transmitted at each wavelength.
  • ⁇ ( ⁇ ) the fraction of light transmitted at each wavelength.
  • the pixel value is specified by the above integral with s( ⁇ ) replaced by ⁇ ( ⁇ ) s( ⁇ ).
  • three color filters Red, Green, and Blue
  • ⁇ r ( ⁇ ), ⁇ g ( ⁇ ), and ⁇ b ( ⁇ ) are used in a color CCD camera with transmittance ⁇ r ( ⁇ ), ⁇ g ( ⁇ ), and ⁇ b ( ⁇ ), respectively.
  • the respective spectral sensitivity functions of the sensor for each primary color is as follows:
  • Equation (5) is a general formula of image formation and pixel value computation of color CCD cameras. More particularly, equation (5) illustrates the relationship among the reflection characteristics of object surface, sensitivity function of CCD pixels, light spectrum of active illumination source, and the signal output of the camera as well as shows the effect of the surface reflectance function r( ⁇ ) plays on the CCD image formation.
  • the RGB value of a CCD pixel is an integral function of three spectral power distribution functions, i.e., spectral composition of active illumination source f( ⁇ ), surface reflectance function r( ⁇ ), and sensitivity function of the CCD camera pixel s( ⁇ ). Even if r( ⁇ ) and s( ⁇ ) are known, it is generally not possible to uniquely determine f( ⁇ ) based on a single set of RGB values to determine the projection angle corresponding to each pixel. Further, in many applications, information about the surface reflectance r( ⁇ ) is not available, making determination of f( ⁇ ) even more unlikely.
  • the inventive system imposes a “single spectral light condition” to the light projector 100 to guarantee the consistency of RGB values corresponding to light spectrum of f( ⁇ ), regardless of the surface reflectance function r( ⁇ ). Under this condition, each individual light sheet projected by the light projector 100 contains only a single wavelength spectral component. More particularly, the spectral composition of the projected light at any single surface point on the object 104 being imaged can be expressed by the following function:
  • m 0 is a variable indicating the magnitude of the light intensity and ⁇ 0 is the wavelength of the single wavelength light impinged on a given surface spot.
  • the normalized RGB values (r,g,b) of a color camera are independent of the intensity of the spectral light m 0 as well as the surface reflectance function r( ⁇ ) as long as they are both non-zero.
  • the spatially varying wavelength light source from the light projector 100 shown in FIG. 2 produces only single spectral light at each surface point on the object being imaged. Due to the property of the delta function of a single wavelength spectral light shown in equation (6), the pixel value p for each pixel in the generated image is:
  • the normalized RGB values, r, g, and b are independent of m 0 and r( ⁇ 0 ).
  • the wavelength information of reflected light captured by a camera 102 in the form of normalized RGB values depends solely on the sensitivity function of the camera 102 , regardless of the light intensity and surface reflectance characteristics of the object 104 being imaged.
  • the sensitivity function of the camera 102 which indicates the responsiveness of a given camera pixel to various wavelengths, can be obtained from the camera manufacturer or through calibration procedures. Based on the normalized RGB data and the known sensitivity function of a camera 102 , one can uniquely determine the corresponding wavelength of the light projected onto the object 104 .
  • the three-dimensional characteristics of the object are then obtained as explained in, for example, U.S. Pat. No. 5,675,407 to obtain the three-dimensional range information of the object 104 . Note that because of the single spectral light condition restriction, the color matching process used to obtain the surface information will work, regardless of whether the object surfaces are pure white, non-white, uniformly colored, mixed colored, or even speckled.

Abstract

A three-dimensional surface profile imaging method and system uses a single spectral light illumination constraint to guarantee consistent RGB values corresponding to a given light spectrum, regardless of the surface reflectance characteristics of the object being imaged. In one embodiment, each light sheet projected onto the object contains only a single wavelength. As a result, the spectral composition of the projected light at any surface point of the object will be independent of the light intensity and the reflectance characteristics of the surface.

Description

REFERENCE TO RELATED APPLICATIONS
This application claims the benefit of U.S. Provisional Appln. No. 60/178,695, filed Jan. 28, 2000, the disclosure of which is incorporated herein by reference in its entirety.
BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention is directed to three-dimensional surface profile imaging, and more particularly to a method and apparatus for three-dimensional imaging that uses color ranging to conduct surface profile measurement.
2. Description of the Related Art
A three dimensional surface profile imaging method and apparatus described in U.S. Pat. No. 5,675,407 (“the '407 patent”), the disclosure of which is incorporated herein by reference in its entirety, conducts imaging by projecting light through a linear variable wavelength filter (LVWF), thereby projecting light having a known, spatially distributed wavelength spectrum on the objects being imaged. The LVWF is a rectangular optical glass plate coated with a color-filtering film that gradually varies in color, (i.e., wavelength). If the color spectrum of a LVWF is within the visible light region, one edge of the filter rectangle may correspond to the shortest visible wavelength (i.e. blue or violet) while the opposite edge may correspond to the longest visible wavelength, (i.e. red). The wavelength of light passing through the coated color-filtering layer is linearly proportional to the distance between the position on the filter glass where the light passes and the blue or red edge. Consequently, the color of the light is directly related to the angle θ, shown in FIG. 1, at which the light leaves the rainbow projector and LVWF.
Referring to FIGS. 1 and 2 in more detail, the imaging method and apparatus is based on the triangulation principle and the relationship between a light projector 100 that projects through the LVWF 101, a camera 102, and the object or scene being imaged 104. As shown in FIG. 1, a triangle is uniquely defined by the angles theta (θ) and alpha (α), and the length of the baseline (B). With known values for θ, α, and B, the distance (i.e., the range R) between the camera 102 and a point Q on the object's surface can be easily calculated. Because the baseline B is predetermined by the relative positions of the light projector 100 and the camera 102, and the value of α can be calculated from the camera's geometry, the key to the triangulation method is to determine the projection angle, θ, from an image captured by the camera 102 and more particularly to determine all θ angles corresponding to all the visible points on an object's surface in order to obtain a full-frame 3D image in one snapshot.
FIG. 2 is a more detailed version of FIG. 1 and illustrates the manner in which all visible points on the object's surface 104 is obtained via the triangulation method. As can be seen in the Figure, the light projector 100 generates a fan beam of light 200. The fan beam 200 is broad spectrum light (i.e., white light) which passes through the LVWF 101 to illuminate one or more three-dimensional objects 104 in the scene with a pattern of light rays possessing a rainbow-like spectrum distribution. The fan beam of light 200 is composed of multiple vertical planes of light 202, or “light sheets”, each plane having a given projection angle and wavelength. Because of the fixed geometric relationship among the light source 100, the lens of the camera 102, and the LVWF 101, there exists a one-to-one correspondence between the projection angle (θ) of the vertical plane of light and the wavelength (λ) of the light ray. Note that although the wavelength variations are shown in FIG. 2 to occur from side to side across the object 104 being imaged, it will be understood by those skilled in the art that the variations in wavelength could also be made from top to bottom across the object 104 or scene being imaged.
The light reflected from the object 104 surface is then detected by the camera 102. If a visible spectrum range LVWF (400-700 nm) is used, the color detected by the camera pixels is determined by the proportion of its primary color Red, Green, and Blue components (RGB). The color spectrum of each pixel has a one-to-one correspondence with the projection angle (θ) of the plane of light due to the fixed geometry of the camera 102 lens and the LVWF 101 characteristics. Therefore, the color of light received by the camera 102 can be used to determine the angle θ at which that light left the light projector 100 through the LVWF 101.
As described above, the angle α is determined by the physical relationship between the camera 102 and the coordinates of each pixel on the camera's imaging plane. The baseline B between the camera's 102 focal point and the center of the cylindrical lens of the light projector 100 is fixed and known. Given the value for angles α and θ, together with the known baseline length B, all necessary information is provided to easily determine the full frame of three-dimensional range values (x,y,z) for any and every visible spot on the surface of the objects 104 seen by the camera 102.
As shown in FIG. 3, given the projection angle θ, the three-dimensional algorithm for determining the (x,y,z) coordinates of any surface spot Q(x,y,z) on a three-dimensional object is given below based on the following triangulation principle: x = B f * ctg θ - u * u , y = B f * ctg θ - u * v , z = B f * ctg θ - u * f ( 1 )
Figure US06556706-20030429-M00001
As a result, the three-dimensional imaging system described above can capture full-frame, high spatial resolution three-dimensional images using a standard camera, such as a charge coupled device camera, in real time without relying on any moving parts. Further, because the imaging system does not rely on a laser, it does not pose any hazard to the eyes when used in clinical applications. Also, because the wavelength of the light projected onto the object surface continuously varies, there is no theoretical limitation on the measurement accuracy that can be achieved by the system. The actual accuracy of a specific system will depend on system implementation and will be affected primarily by limiting factors such as the optical system design, the quality and resolution of the camera, light spectral emission of the light source projector; noise level and resolution of the frame grabber, calibration algorithms, and the three-dimensional imaging processing algorithms.
To avoid allowing the surface color of the object being imaged from affecting the imaging results, the system may obtain an image of the object under normal light conditions before projecting the filtered light onto the object. The image obtained under normal light conditions is then subtracted from the image obtained under LVWF light conditions to eliminate the effects of the object color on the image.
Even when the system compensates for the color of the object, however, the consistency of the spectral power distribution and the RGB value of each pixel may vary when light is projected onto the object through the LVWF based on the reflection characteristics of the object's surface, particularly if the object is not white and/or not uniformly colored.
There is a need for a surface profile imaging method and apparatus that is able to generate consistent RGB values regardless of the reflection characteristics of the surface being imaged.
SUMMARY OF THE INVENTION
Accordingly, the present invention is directed to a method and apparatus for three-dimensional surface imaging that avoids variations in the RGB value of each pixel due to the reflection characteristics of the object's surface. More particularly, a light source in the system illuminates an object or scene with a light pattern having a spatially varying wavelength and composed of at least one light plane. The light plane corresponds to at least one angle at which the light of that wavelength is emitted and contains only a single spectral component.
By imposing a single spectral light condition on the light source, the RGB values of each pixel will be independent of the light intensity of the light source and the reflectance characteristics of the object or scene being imaged. As a result, any color matching function that is conducted to link the wavelength of the light projected on the object or scene at a given point and that point's position will be consistent, regardless of the color and/or reflectance characteristics of the object's or scene's surface.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a simplified diagram illustrating a triangulation principle used in the present invention;
FIG. 2 is a representative diagram of the components used by the inventive system; and
FIG. 3 is a plan view of the system shown in FIG. 2.
DESCRIPTION OF THE PREFERRED EMBODIMENT
Generally, the inventive system and method system ensures that the RGB values corresponding to a given spectral composition f(λ) are consistent, regardless of the reflectance characteristics of the surface being imaged. By way of background, the invention recognizes that the spectral composition f(λ) of the light source 100 is a parameter that is controllable in the system shown in FIG. 2. Because this parameter is controllable, imposing a restrictive condition on f(λ) can make the RGB values consistent for f(λ), regardless of r(λ) and s(λ). This will be explained in greater detail below.
In a monochrome camera, the pixel values are computed from the integral of the spectral power distribution (SPD) of the light measured by a charge coupled device (CCD) element in the camera 102. If the camera 102 receives light directly from a light source, a given pixel in the CCD element of the camera 102 generates output value p by summing the amount of light f(λ) received by the pixel at each wavelength, weighted by the responsiveness of the camera pixel to a given wavelength, s(λ). That is: p = λ min λ max f ( λ ) s ( λ ) λ ( 2 )
Figure US06556706-20030429-M00002
where f(λ) is the spectral composition (spectral power distribution) of illumination source; s(λ) is the spectral sensitivity function of the sensor; and λ max and λ min are the upper and lower bounds of the wavelengths visible to CCD pixels in the CCD elements.
When the camera 102 is used to obtain images of real world objects, where the light source position may not be controllable, the reflectance characteristics of the surface of the objects will affect the pixel values because the camera 102 will not receive light directly from the light source in most real-world cases. The pixel value p in this case is a integral of the amount of light f(λ) received by the pixel of the camera at each wavelength, weighted by the surface reflectance function of the object, r(λ), weighted further by the responsiveness of the camera pixel to various wavelength, s(λ). That is: p = λ min λ max f ( λ ) r ( λ ) s ( λ ) λ ( 3 )
Figure US06556706-20030429-M00003
where r(λ) is the surface spectral reflectance function of the object in the scene 104.
If the camera 102 is a CCD camera rather than a monochrome camera, color filters are interposed between the incoming illumination and the pixel element of the camera 102. Each filter has a transmittance function τ(λ), specifying the fraction of light transmitted at each wavelength. Thus, the pixel value is specified by the above integral with s(λ) replaced by τ(λ) s(λ). Typically, three color filters (Red, Green, and Blue) are used in a color CCD camera with transmittance τr(λ), τg(λ), and τb(λ), respectively. More particularly, the respective spectral sensitivity functions of the sensor for each primary color is as follows:
s r(λ)=τr(λ)s(λ), s g(λ)=τg(λ)s(λ), s b(λ)=τb(λ)s(λ)  (4)
As a result, the RGB value of each pixel of a color camera can be computed by the expression: [ R G B ] = [ λ min λ max f ( λ ) r ( λ ) s r ( λ ) λ λ min λ max f ( λ ) r ( λ ) s g ( λ ) λ λ min λ max f ( λ ) r ( λ ) s b ( λ ) λ ] ( 5 )
Figure US06556706-20030429-M00004
Equation (5) is a general formula of image formation and pixel value computation of color CCD cameras. More particularly, equation (5) illustrates the relationship among the reflection characteristics of object surface, sensitivity function of CCD pixels, light spectrum of active illumination source, and the signal output of the camera as well as shows the effect of the surface reflectance function r(λ) plays on the CCD image formation.
As can be seen in equation (5), the RGB value of a CCD pixel is an integral function of three spectral power distribution functions, i.e., spectral composition of active illumination source f(λ), surface reflectance function r(λ), and sensitivity function of the CCD camera pixel s(λ). Even if r(λ) and s(λ) are known, it is generally not possible to uniquely determine f(λ) based on a single set of RGB values to determine the projection angle corresponding to each pixel. Further, in many applications, information about the surface reflectance r(λ) is not available, making determination of f(λ) even more unlikely.
Note, however, that it is not necessary when using a LVWF 101 in the imaging system to recover explicitly the function form of f(λ) based on RGB values, due to the color matching scheme used to determine the projection angle. Instead, the detected light spectrum is compared with values pre-stored in a look-up table in the camera's processor or in a separate processor coupled to the camera to get a corresponding projection angle. In this case, the spectrum of projected light is only used as an information carrier for the projection angle of the light sheet. As long as the CCD camera can provide the same RGB values for the same projected light illumination (with various reflectance functions r(λ)), the specific form of the function f(λ) is not important. Instead, consistency of RGB values corresponding to light spectrum of f(λ), regardless of the surface reflectance function r(λ), has greater importance because the consistency facilitates color matching.
With both f(λ) and r(λ) being freely variable, however, the consistency of RGB values corresponding to f(λ) is difficult to achieve, thereby making color matching difficult, particularly because differences in the surface characteristics (e.g., surface color, reflectivity of the surface material, non-uniform surface color, etc.) will change the RGB values.
To restrict the spectral composition f(λ), the inventive system imposes a “single spectral light condition” to the light projector 100 to guarantee the consistency of RGB values corresponding to light spectrum of f(λ), regardless of the surface reflectance function r(λ). Under this condition, each individual light sheet projected by the light projector 100 contains only a single wavelength spectral component. More particularly, the spectral composition of the projected light at any single surface point on the object 104 being imaged can be expressed by the following function:
f(λ)=m 0δ(λ−λ0)  (6)
where m0 is a variable indicating the magnitude of the light intensity and λ0 is the wavelength of the single wavelength light impinged on a given surface spot.
Under this single spectral light illumination condition, the normalized RGB values (r,g,b) of a color camera are independent of the intensity of the spectral light m0 as well as the surface reflectance function r(λ) as long as they are both non-zero. More particularly, under the single spectral light illumination condition, the spatially varying wavelength light source from the light projector 100 shown in FIG. 2 produces only single spectral light at each surface point on the object being imaged. Due to the property of the delta function of a single wavelength spectral light shown in equation (6), the pixel value p for each pixel in the generated image is:
p=m 0 r0)s0)  (7)
where λ0 is the wavelength of the single wavelength light impinged on the surface spot observed by the CCD pixel. Based on equation (7), the R, G, and B values of each pixel therefore can be expressed as: [ R G B ] = m 0 r ( λ 0 ) [ s r ( λ 0 ) s g ( λ 0 ) s b ( λ 0 ) ] ( 8 )
Figure US06556706-20030429-M00005
The wavelength information of projected light is embedded in the normalized RGB values of each pixel. Assuming that both m0 and r(λ0) are non-zero, the normalized RGB values are expressed as follows: [ r g b ] = 1 R 2 + G 2 + B 2 [ R G B ] = 1 s r ( λ 0 ) 2 + s g ( λ 0 ) 2 + s b ( λ 0 ) 2 [ s r ( λ 0 ) s g ( λ 0 ) s b ( λ 0 ) ] ( 9 )
Figure US06556706-20030429-M00006
As can be seen in equation (9), the normalized RGB values, r, g, and b are independent of m0 and r(λ0).
Under the single spectrum light condition, then, the wavelength information of reflected light captured by a camera 102 in the form of normalized RGB values depends solely on the sensitivity function of the camera 102, regardless of the light intensity and surface reflectance characteristics of the object 104 being imaged. The sensitivity function of the camera 102, which indicates the responsiveness of a given camera pixel to various wavelengths, can be obtained from the camera manufacturer or through calibration procedures. Based on the normalized RGB data and the known sensitivity function of a camera 102, one can uniquely determine the corresponding wavelength of the light projected onto the object 104. The three-dimensional characteristics of the object are then obtained as explained in, for example, U.S. Pat. No. 5,675,407 to obtain the three-dimensional range information of the object 104. Note that because of the single spectral light condition restriction, the color matching process used to obtain the surface information will work, regardless of whether the object surfaces are pure white, non-white, uniformly colored, mixed colored, or even speckled.
Note that although the above description focuses on using visible light for conducting the three-dimensional surface imaging, infrared (IR) or ultra-violet (UV) light sources with suitable wavelength sensitive detectors can also be used for special applications with minimal modification of the imaging system. These applications will not be specifically addressed herein, but will be readily practicable by those skilled in the art based on the principles disclosed herein.
While the invention has been specifically described in connection with certain specific embodiments thereof, it is to be understood that this is by way of illustration and not of limitation, and the scope of the appended claims should be construed as broadly as the prior art will permit.

Claims (16)

What is claimed is:
1. A device for creating a three-dimensional profile of an object or scene being imaged, the device comprising:
a light source for illuminating said object or scene with a light pattern having at least one light plane, wherein said light pattern varies in wavelength spatially across said object or scene, and wherein said at least one light plane corresponds to at least one angle at which light of that wavelength is emitted by said light source and contains only a single spectral component;
a camera for imaging said object or scene as illuminated with said light pattern; and
a processor for calculating a distance to a point on said object or in said scene based on a baseline distance between said light source and said camera, an angle between said camera and said baseline, and an angle at which light striking the point is emitted by said light source as determined from red/green/blue (RGB) values corresponding to the wavelength of the light striking the point.
2. The device of claim 1, wherein the single spectral component is based on a light intensity value of the light source and the wavelength emitted by said light source.
3. The device of claim 2, wherein the RGB values depend on a sensitivity function of the camera and are independent of the light intensity of the light source and a surface reflectance of the object or scene.
4. The device of claim 1, wherein said light pattern varies over a visible light spectrum.
5. The device of claim 1, wherein said light pattern varies over an infrared light spectrum.
6. The device of claim 1, wherein said light pattern varies over an ultraviolet light spectrum.
7. The device of claim 1, wherein the processor calculates the distance to the point by triangulation of the baseline distance between said light source and said camera, the angle between said camera and said baseline, and the angle at which light striking the point is emitted by said light source.
8. A device for creating a three-dimensional profile of an object or scene being imaged, the device comprising:
a light source for illuminating said object or scene with a light pattern having a plurality of light planes, wherein said light pattern varies in wavelength spatially across said object or scene, and wherein each of said plurality of light planes corresponds to at least one angle at which light of that wavelength is emitted by said light source and contains only a single spectral component, the single spectral component being based on a light intensity value of the light source and the wavelength emitted by said light source for one of said light planes;
a camera for imaging said object or scene as illuminated with said light pattern; and
a processor for calculating a distance to a point on said object or in said scene using triangulation based on a baseline distance between said light source and said camera, an angle between said camera and said baseline, and an angle at which light striking the point is emitted by said light source as determined from red/green/blue (RGB) values corresponding to the wavelength of the light striking the point, wherein the RGB values depend on a sensitivity function of the camera and are independent of the light intensity of the light source and a surface reflectance of the object or scene.
9. The device of claim 8, wherein said light pattern varies over a visible light spectrum.
10. The device of claim 8, wherein said light pattern varies over an infrared light spectrum.
11. The device of claim 8, wherein said light pattern varies over an ultraviolet light spectrum.
12. A method for creating a three-dimensional profile of an object or scene being imaged, the method comprising the steps of:
illuminating the object or scene with a light pattern having a plurality of light planes, wherein said light pattern varies in wavelength spatially across said object or scene, and wherein said at least one light plane corresponds to at least one angle at which light of that wavelength is emitted and contains only a single spectral component;
imaging said object or scene as illuminated in said illuminating step;
calculating a distance to a point on said object or in said scene using triangulation based on a baseline distance between said light source and said camera, an angle between said camera and said baseline, and an angle at which light striking the point is emitted by said light source as determined from red/green/blue (RGB) values corresponding to the wavelength of the light striking the point.
13. The method of claim 12, wherein the RGB values depend on a sensitivity function of the camera and are independent of the light intensity of the light source and a surface reflectance of the object or scene.
14. The method of claim 12, wherein the illuminating step varies the light pattern over a visible light spectrum.
15. The method of claim 12, wherein the illuminating step varies the light pattern over an infrared light spectrum.
16. The method of claim 12, wherein the illuminating step varies the light pattern over an ultraviolet light spectrum.
US09/770,124 2000-01-28 2001-01-26 Three-dimensional surface profile imaging method and apparatus using single spectral light condition Expired - Fee Related US6556706B1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
US09/770,124 US6556706B1 (en) 2000-01-28 2001-01-26 Three-dimensional surface profile imaging method and apparatus using single spectral light condition
EP01946201A EP1354176A1 (en) 2001-01-26 2001-06-11 Three-dimensional surface profile imaging method and apparatus using single spectral light condition
AU2001268281A AU2001268281A1 (en) 2001-01-26 2001-06-11 Three-dimensional surface profile imaging method and apparatus using single spectral light condition
PCT/US2001/018644 WO2002059545A1 (en) 2001-01-26 2001-06-11 Three-dimensional surface profile imaging method and apparatus using single spectral light condition

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US17869500P 2000-01-28 2000-01-28
US09/770,124 US6556706B1 (en) 2000-01-28 2001-01-26 Three-dimensional surface profile imaging method and apparatus using single spectral light condition

Publications (1)

Publication Number Publication Date
US6556706B1 true US6556706B1 (en) 2003-04-29

Family

ID=25087555

Family Applications (1)

Application Number Title Priority Date Filing Date
US09/770,124 Expired - Fee Related US6556706B1 (en) 2000-01-28 2001-01-26 Three-dimensional surface profile imaging method and apparatus using single spectral light condition

Country Status (4)

Country Link
US (1) US6556706B1 (en)
EP (1) EP1354176A1 (en)
AU (1) AU2001268281A1 (en)
WO (1) WO2002059545A1 (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040179729A1 (en) * 2003-03-13 2004-09-16 Minolta Co., Ltd. Measurement system
US20040212811A1 (en) * 2001-12-07 2004-10-28 Takeo Iwasaki Three-dimensional shape and color detecting apparatus
US20040218788A1 (en) * 2003-01-31 2004-11-04 Geng Z. Jason Three-dimensional ear biometrics system and method
US20060127854A1 (en) * 2004-12-14 2006-06-15 Huafeng Wen Image based dentition record digitization
US20060127852A1 (en) * 2004-12-14 2006-06-15 Huafeng Wen Image based orthodontic treatment viewing system
US20060127836A1 (en) * 2004-12-14 2006-06-15 Huafeng Wen Tooth movement tracking system
WO2007113377A1 (en) 2006-04-05 2007-10-11 Seppaelae Kari Method for measuring/recognizing a shape
CN100387065C (en) * 2003-07-07 2008-05-07 财团法人工业技术研究院 Three-dimensional color information acquisition method and device therefor
US20090066929A1 (en) * 2006-01-08 2009-03-12 Hermann Tropf Creation of a Range Image
US20090185800A1 (en) * 2008-01-23 2009-07-23 Sungkyunkwan University Foundation For Corporate Collaboration Method and system for determining optimal exposure of structured light based 3d camera
US20110050859A1 (en) * 2009-09-03 2011-03-03 Technion Research & Development Foundation Ltd. Devices and methods of generating three dimensional (3d) colored models
CN102322822A (en) * 2011-08-08 2012-01-18 西安交通大学 Three-dimensional measurement method for triple-frequency color fringe projection
US20120229606A1 (en) * 2009-11-04 2012-09-13 Technologies Numetrix Inc. Device and method for obtaining three-dimensional object surface data
CN102102980B (en) * 2009-12-21 2013-06-12 重庆工商大学 Method and device for measuring axial displacement by using sing primary color contrast as characteristic
US20140374603A1 (en) * 2013-06-20 2014-12-25 The University Of North Carolina At Charlotte Profilometry systems and methods based on absorption and optical frequency conversion
US8928892B2 (en) 2009-03-04 2015-01-06 Elie Meimoun Wavefront analysis inspection apparatus and method
US8941825B2 (en) 2013-03-15 2015-01-27 Owens-Brockway Glass Container Inc. Container inspection
US20160307382A1 (en) * 2015-04-10 2016-10-20 Google Inc. Method and system for optical user recognition
US9886759B2 (en) 2013-10-21 2018-02-06 National Taiwan University Of Science And Technology Method and system for three-dimensional data acquisition
US20180087898A1 (en) * 2015-05-29 2018-03-29 Nippon Steel & Sumitomo Metal Corporation Shape inspection apparatus for metallic body and shape inspection method for metallic body

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
SG11201609800SA (en) 2012-12-10 2016-12-29 Dirtt Environmental Solutions Efficient lighting effects in design software
US9355197B2 (en) 2013-01-25 2016-05-31 Dirtt Environmental Solutions, Ltd Real-time depth of field effects within design software
EP2951785B1 (en) * 2013-01-31 2019-12-11 Dirtt Environmental Solutions, Ltd. Method and system for efficient modeling of specular reflection
US9245381B2 (en) 2013-01-31 2016-01-26 Ice Edge Business Solutions, Ltd Visual distortion effects through translucent structures in design software
EP2956873A4 (en) 2013-05-31 2016-11-02 Ice Edge Business Solutions Ltd Associating computer-executable objects with three-dimensional spaces within an architectural design environment
EP3152373A4 (en) 2014-06-09 2018-05-02 Dirtt Environmental Solutions Inc. Associating computer-executable objects with timber frames within an architectural design environment
US10659699B2 (en) * 2014-07-09 2020-05-19 Asm Technology Singapore Pte Ltd Apparatus and method for reconstructing a three-dimensional profile of a target surface

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5090811A (en) * 1989-05-31 1992-02-25 General Electric Company Optical radius gauge
US5471303A (en) * 1994-04-29 1995-11-28 Wyko Corporation Combination of white-light scanning and phase-shifting interferometry for surface profile measurements
US6028672A (en) 1996-09-30 2000-02-22 Zheng J. Geng High speed three dimensional imaging method
US6147760A (en) 1994-08-30 2000-11-14 Geng; Zheng Jason High speed three dimensional imaging method
WO2000070303A1 (en) 1999-05-14 2000-11-23 3Dmetrics, Incorporated Color structured light 3d-imaging system
US6252623B1 (en) * 1998-05-15 2001-06-26 3Dmetrics, Incorporated Three dimensional imaging system

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5090811A (en) * 1989-05-31 1992-02-25 General Electric Company Optical radius gauge
US5471303A (en) * 1994-04-29 1995-11-28 Wyko Corporation Combination of white-light scanning and phase-shifting interferometry for surface profile measurements
US6147760A (en) 1994-08-30 2000-11-14 Geng; Zheng Jason High speed three dimensional imaging method
US6028672A (en) 1996-09-30 2000-02-22 Zheng J. Geng High speed three dimensional imaging method
US6252623B1 (en) * 1998-05-15 2001-06-26 3Dmetrics, Incorporated Three dimensional imaging system
WO2000070303A1 (en) 1999-05-14 2000-11-23 3Dmetrics, Incorporated Color structured light 3d-imaging system

Cited By (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040212811A1 (en) * 2001-12-07 2004-10-28 Takeo Iwasaki Three-dimensional shape and color detecting apparatus
US6927862B2 (en) * 2001-12-07 2005-08-09 Brother Kogyo Kabushiki Kaisha Three-dimensional shape and color detecting apparatus
US20060140453A1 (en) * 2003-01-31 2006-06-29 Geng Z J Three-dimensional ear biometrics system and method
US20040218788A1 (en) * 2003-01-31 2004-11-04 Geng Z. Jason Three-dimensional ear biometrics system and method
US7826643B2 (en) 2003-01-31 2010-11-02 Technest Holdings, Inc. Three-dimensional ear biometrics system and method
US7065232B2 (en) 2003-01-31 2006-06-20 Genex Technologies, Inc. Three-dimensional ear biometrics system and method
US20040179729A1 (en) * 2003-03-13 2004-09-16 Minolta Co., Ltd. Measurement system
CN100387065C (en) * 2003-07-07 2008-05-07 财团法人工业技术研究院 Three-dimensional color information acquisition method and device therefor
US20060127854A1 (en) * 2004-12-14 2006-06-15 Huafeng Wen Image based dentition record digitization
US20070141534A1 (en) * 2004-12-14 2007-06-21 Huafeng Wen Image-based orthodontic treatment viewing system
US20060127836A1 (en) * 2004-12-14 2006-06-15 Huafeng Wen Tooth movement tracking system
US20060127852A1 (en) * 2004-12-14 2006-06-15 Huafeng Wen Image based orthodontic treatment viewing system
US20090066929A1 (en) * 2006-01-08 2009-03-12 Hermann Tropf Creation of a Range Image
US7804585B2 (en) 2006-01-08 2010-09-28 Visiontools Bildanalyse Systeme Gmbh Creation of a range image
WO2007113377A1 (en) 2006-04-05 2007-10-11 Seppaelae Kari Method for measuring/recognizing a shape
US20090185800A1 (en) * 2008-01-23 2009-07-23 Sungkyunkwan University Foundation For Corporate Collaboration Method and system for determining optimal exposure of structured light based 3d camera
US7957639B2 (en) * 2008-01-23 2011-06-07 Sungkyunkwan University Foundation For Corporate Collaboration Method and system for determining optimal exposure of structured light based 3D camera
USRE48595E1 (en) * 2008-01-23 2021-06-15 Sungkyunkwan University Foundation For Corporate Collaboration Method and system for determining optimal exposure of structured light based 3D camera
US8928892B2 (en) 2009-03-04 2015-01-06 Elie Meimoun Wavefront analysis inspection apparatus and method
US20110050859A1 (en) * 2009-09-03 2011-03-03 Technion Research & Development Foundation Ltd. Devices and methods of generating three dimensional (3d) colored models
US8908958B2 (en) * 2009-09-03 2014-12-09 Ron Kimmel Devices and methods of generating three dimensional (3D) colored models
US20120229606A1 (en) * 2009-11-04 2012-09-13 Technologies Numetrix Inc. Device and method for obtaining three-dimensional object surface data
US8520058B2 (en) * 2009-11-04 2013-08-27 Technologies Numetrix Inc. Device and method for obtaining three-dimensional object surface data
US8462357B2 (en) 2009-11-04 2013-06-11 Technologies Numetrix Inc. Device and method for obtaining three-dimensional object surface data
CN102102980B (en) * 2009-12-21 2013-06-12 重庆工商大学 Method and device for measuring axial displacement by using sing primary color contrast as characteristic
CN102322822A (en) * 2011-08-08 2012-01-18 西安交通大学 Three-dimensional measurement method for triple-frequency color fringe projection
CN102322822B (en) * 2011-08-08 2013-04-17 西安交通大学 Three-dimensional measurement method for triple-frequency color fringe projection
US8941825B2 (en) 2013-03-15 2015-01-27 Owens-Brockway Glass Container Inc. Container inspection
US9250186B2 (en) * 2013-06-20 2016-02-02 The University Of North Carolina At Charlotte Profilometry systems and methods based on absorption and optical frequency conversion
US20140374603A1 (en) * 2013-06-20 2014-12-25 The University Of North Carolina At Charlotte Profilometry systems and methods based on absorption and optical frequency conversion
US9886759B2 (en) 2013-10-21 2018-02-06 National Taiwan University Of Science And Technology Method and system for three-dimensional data acquisition
US20160307382A1 (en) * 2015-04-10 2016-10-20 Google Inc. Method and system for optical user recognition
US9984519B2 (en) * 2015-04-10 2018-05-29 Google Llc Method and system for optical user recognition
US20180087898A1 (en) * 2015-05-29 2018-03-29 Nippon Steel & Sumitomo Metal Corporation Shape inspection apparatus for metallic body and shape inspection method for metallic body
US9970750B2 (en) * 2015-05-29 2018-05-15 Nippon Steel & Sumitomo Metal Corporation Shape inspection apparatus for metallic body and shape inspection method for metallic body

Also Published As

Publication number Publication date
WO2002059545A8 (en) 2002-08-29
EP1354176A1 (en) 2003-10-22
WO2002059545A1 (en) 2002-08-01
AU2001268281A1 (en) 2002-08-06

Similar Documents

Publication Publication Date Title
US6556706B1 (en) Three-dimensional surface profile imaging method and apparatus using single spectral light condition
US6147760A (en) High speed three dimensional imaging method
US8244023B2 (en) Shape measuring device and shape measuring method
US6700669B1 (en) Method and system for three-dimensional imaging using light pattern having multiple sub-patterns
US8199335B2 (en) Three-dimensional shape measuring apparatus, three-dimensional shape measuring method, three-dimensional shape measuring program, and recording medium
JP5396012B2 (en) System that automatically corrects the image before projection
WO2019133631A2 (en) System and method of imaging using multiple illumination pulses
JP3417222B2 (en) Real-time range finder
US6765606B1 (en) Three dimension imaging by dual wavelength triangulation
US20100114265A1 (en) Device for irradiating an object, in particular the human skin, with uv light
KR101824328B1 (en) 3-D scanner and the scanning method using the chromatic aberration
US11902494B2 (en) System and method for glint reduction
CN113454419A (en) Detector having a projector for illuminating at least one object
US20150098092A1 (en) Device and Method For the Simultaneous Three-Dimensional Measurement of Surfaces With Several Wavelengths
US11792383B2 (en) Method and system for reducing returns from retro-reflections in active illumination system
JP2011510315A (en) Spatial colorimetric measuring device and spatial colorimetric measuring method for three-dimensional objects
TWI801637B (en) Infrared pre-flash for camera
JP3414624B2 (en) Real-time range finder
JPS62223634A (en) Color deciding device
JP7005175B2 (en) Distance measuring device, distance measuring method and imaging device
JP2023169165A (en) Computer, program, and method
JPH07117475B2 (en) Lighting characteristic evaluation device for light source unit
US20130063408A1 (en) Projection device, which comprises a projector, a projection surface, and a data processing system, and method for operating said projection device
JP7441380B2 (en) distance measuring device
JPH0427813A (en) Method and apparatus for obtaining distance image

Legal Events

Date Code Title Description
AS Assignment

Owner name: GENEX TECHNOLOGIES, INC., MARYLAND

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GENG, ZHENG JASON;REEL/FRAME:015778/0024

Effective date: 20050211

AS Assignment

Owner name: SILICON VALLEY BANK, CALIFORNIA

Free format text: SECURITY AGREEMENT;ASSIGNORS:TECHNEST HOLDINGS, INC.;E-OIR TECHNOLOGIES, INC.;GENEX TECHNOLOGIES INCORPORATED;REEL/FRAME:018148/0292

Effective date: 20060804

FPAY Fee payment

Year of fee payment: 4

AS Assignment

Owner name: TECHNEST HOLDINGS, INC., MASSACHUSETTS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GENEX TECHNOLOGIES, INC.;REEL/FRAME:019781/0010

Effective date: 20070406

AS Assignment

Owner name: E-OIR TECHNOLOGIES, INC., VIRGINIA

Free format text: RELEASE;ASSIGNOR:SILICON VALLEY BANK;REEL/FRAME:020462/0938

Effective date: 20080124

Owner name: GENEX TECHNOLOGIES INCORPORATED, VIRGINIA

Free format text: RELEASE;ASSIGNOR:SILICON VALLEY BANK;REEL/FRAME:020462/0938

Effective date: 20080124

Owner name: TECHNEST HOLDINGS, INC., VIRGINIA

Free format text: RELEASE;ASSIGNOR:SILICON VALLEY BANK;REEL/FRAME:020462/0938

Effective date: 20080124

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20110429