US20020186878A1 - System and method for multiple image analysis - Google Patents

System and method for multiple image analysis Download PDF

Info

Publication number
US20020186878A1
US20020186878A1 US09/876,795 US87679501A US2002186878A1 US 20020186878 A1 US20020186878 A1 US 20020186878A1 US 87679501 A US87679501 A US 87679501A US 2002186878 A1 US2002186878 A1 US 2002186878A1
Authority
US
United States
Prior art keywords
image data
component
data
receiving
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US09/876,795
Inventor
Tan Hoon
Sreenivas Rao
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
ASTI HOLDINGS Inc
ASTI Holdings Ltd
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US09/876,795 priority Critical patent/US20020186878A1/en
Assigned to ASTI HOLDINGS, INC. reassignment ASTI HOLDINGS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RAO, SREENIVAS, HOON, TAN SEOW
Priority to PCT/IB2002/002050 priority patent/WO2002099403A1/en
Assigned to ASTI HOLDINGS LIMITED reassignment ASTI HOLDINGS LIMITED CORRECTIVE ASSIGNMENT TO CORRECT THE RECEIVING PARTY ADDRESS PREVIOUSLY RECORDED ON REEL 012429 FRAME 0766 ASSIGNOR HEREBY CONFIRMS THE ASSIGNMENT OF THE ENTIRE INTEREST. Assignors: SREENIVAS, RAO, HOON, TAN SEOW
Publication of US20020186878A1 publication Critical patent/US20020186878A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/84Systems specially adapted for particular applications
    • G01N21/88Investigating the presence of flaws or contamination
    • G01N21/95Investigating the presence of flaws or contamination characterised by the material or shape of the object to be examined
    • G01N21/956Inspecting patterns on the surface of objects
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/24Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
    • G01B11/245Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures using a plurality of fixed, simultaneously operating transducers
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0004Industrial image inspection
    • G06T7/001Industrial image inspection using an image reference approach
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/586Depth or shape recovery from multiple images from multiple light sources, e.g. photometric stereo
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/62Analysis of geometric attributes of area, perimeter, diameter or volume
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30108Industrial image inspection
    • G06T2207/30148Semiconductor; IC; Wafer

Definitions

  • the present invention pertains to the field of semiconductor devices, and more particularly to a system and method for inspecting semiconductor devices that uses multiple two-dimensional images to generate third dimension data.
  • Image data analysis systems for inspecting semiconductor components are known in the art. Such image data analysis systems attempt to determine the state of the semiconducting component or other inspected components by analyzing image data, which is typically comprised of an N ⁇ M array of picture elements or “pixels.” The brightness value of each pixel of a test image is typically compared to the brightness element of a corresponding pixel of a reference image, and the comparison data is analyzed to determine whether or not unacceptable defects exist on the semiconducting device, component or other object being inspected. For example, image data analysis is used to determine whether the variation in the dimensions of an element of the component exceed allowable tolerances for such dimensions.
  • One drawback with known image data inspection systems is the difficulty in determining the three-dimensional nature of elements. Such image data is typically taken from a single angle, such that any three-dimensional aspects or flaws may be difficult to detect.
  • a common method for determining the three-dimensional aspects of a semiconductor device or component that is being inspected is to use a laser beam to trace a line, and to determine when the line varies from a straight line, where such variations are then correlated to defects in the semiconducting device or component.
  • the semiconducting device or component contains a large number of elements, it is necessary to trace a laser line through each of the elements, which can require movement of the component to a number of different locations. Likewise, it is possible that the laser drawn line may not lie on a defect, such that the defect could be missed.
  • a system and method for multiple image analysis are provided that use image data generated by illuminating a component from two or more lighting angles, which allows three-dimensional aspects of the component to be determined.
  • a system for analyzing multiple images is provided, such as to locate defects in a test component.
  • the system includes a first light source, such as one that emits blue light, and a second light source, such as one that emits red light.
  • the system also includes a camera, where the camera and the light sources are focused on an area where a test piece is to be placed.
  • a multiple image processor is connected to the first light source, the second light source, and the camera.
  • the multiple image processor causes the first light source and the second light source to turn on, such as in sequence, and also causes the camera to generate two or more sets of image data, such as one set when each of the light sources is illuminated, through the use of filters or tuned pixels, or otherwise.
  • the present invention provides many important technical advantages.
  • One important technical advantage of the present invention is a system and method for multiple image analysis that uses two or more sets of image data to analyze a component. Each set of image data is obtained when the component is illuminated by a light source having a different lighting angle, which creates shaded areas that can be analyzed to determine whether they indicate the existence of damage or unacceptable dimensional variations on the component.
  • FIG. 1 is a diagram of a system for performing multiple image analysis in accordance with an exemplary embodiment of the present invention
  • FIGS. 2A, 2B, and 2 C show an exemplary undamaged element and corresponding bright and shaded regions generated by illumination from light sources;
  • FIGS. 3A, 3B, and 3 C show an exemplary damaged element and corresponding bright and shaded regions generated by illumination from light sources
  • FIG. 4 is a diagram of a system for processing image data from multiple images in accordance with an exemplary embodiment of the present invention
  • FIG. 5 is a flowchart of a method for analyzing image data from multiple images in accordance with an exemplary embodiment of the present invention
  • FIG. 6 is a flowchart of a method for analyzing image data in accordance with an exemplary embodiment of the present invention.
  • FIG. 7 is a flowchart of a method for performing image data analysis for multiple images in accordance with an exemplary embodiment of the present invention.
  • FIG. 1 is a diagram of a system 100 for performing multiple image analysis in accordance with an exemplary embodiment of the present invention.
  • System 100 allows three-dimensional aspects of an inspected device or component to be determined from images obtained from two or more different viewing angles.
  • System 100 includes multiple image processor 102 , which can be implemented in hardware, software, or a suitable combination of hardware and software, and which can be one or more software systems operating on a general purpose processing platform.
  • a software system can include one or more objects, agents, subroutines, lines of code, threads, two or more lines of code or other suitable software structures operating in two or more separate software applications, or other suitable software structures, and can operate on two or more different processors, or other suitable configurations of processors.
  • a software system can include one or more lines of code or other software structures operating in a general purpose software application, such as an operating system, and one or more lines of code or other suitable software structures operating in a specific purpose software application.
  • Coupled to light sources 104 a and 104 b can include a physical connection (such as a copper conductor), a virtual connection (such as through one or more randomly assigned data memory locations of a data memory device), a logical connection (such as through one or more logical gates of a semiconducting device), a wireless connection, other suitable connections, or a suitable combination of such connections.
  • systems and components are coupled to other systems and components through intervening systems and components, such as through an operating system of a general purpose processor platform.
  • Multiple image processor 102 is also coupled to camera 106 .
  • Camera 106 can be a charge coupled device (CCD), a CMOS imaging device, or other suitable imaging devices that are focused on a component 108 having a plurality of elements 110 .
  • Light sources 104 a and 104 b are also focused on component 108 , and illuminate component 108 from different angles as shown in FIG. 1.
  • the light illuminating component 108 from light source 104 a will create shaded regions that are different from the shaded regions created by light illuminating component 108 from light source 104 b .
  • Additional light sources can be used where suitable to create additional shaded regions.
  • Camera 106 is used to record image data of component 108 while it is being illuminated by light sources 104 a and 104 b .
  • camera 106 is controlled by multiple image processor 102 to store a first set of image data of component 108 when light source 104 a is on, and to store a second set of image data when light source 104 b is on.
  • camera 106 can store image data when both of light sources 104 a and 104 b are on, such as when the light sources use different frequencies of light.
  • camera 106 can record image data according to the frequency of the light that creates the image, such as by including one or more light filters, two or more sets of pixels that are tuned to received predetermined frequencies of light, or to otherwise differentiate between light illuminated from light sources 104 a and 104 b , such that multiple sets of image data can be concurrently gathered.
  • a component 108 is placed in the focal area of light sources 104 a and 104 b and camera 106 for inspection.
  • Multiple image processor 102 then causes component 108 to be illuminated and causes camera 106 to produce image data, such as by generating an N ⁇ M array of pixels of image data, which can then be stored by multiple image processor 102 or other suitable storage systems or devices. Because of the angular difference between light sources 104 a and 104 b relative to component 108 , shaded regions are generated from elements 110 .
  • Multiple image processor 102 can analyze these shaded regions to determine whether they are indicative of any damage or defects to component 108 , elements 110 , or other suitable indications.
  • multiple image processor 102 can determine whether three-dimensional defects or other variations in component 108 or elements 110 exist. For example, if one of elements 110 is damaged, then the shaded regions generated by that element 110 when it is illuminated by light sources 104 a and 104 b will vary from the shaded regions generated for undamaged reference images. Furthermore, the variations in pixel brightness between corresponding pixels of the test image data and the reference image data, as illuminated by light sources 104 a and 104 b , can also be analyzed to generate an approximation of differences in height, dimensions, or other data that can be used to approximate a three-dimensional analysis.
  • FIGS. 2A, 2B, and 2 C show an exemplary undamaged element 110 and corresponding bright and shaded regions generated by illumination from light sources 104 a and 104 b (not explicitly shown).
  • FIG. 2A shows an exemplary undamaged element 110 , which is semi-spherical in shape.
  • the circular outline of element 110 as viewed from overhead is shown with an illuminated region and a shaded region corresponding to the shadow generated by light source 104 a .
  • the shaded region generates a distinctive pattern which is indicative of a spherical configuration of element 110 .
  • the shaded region of element 110 is on the opposite face, as a result of the location of light source 104 b .
  • the shaded regions generated shown in FIGS. 2 b and 2 c can be used as a reference for an undamaged element 110 .
  • the differences in pixel brightness data between FIG. 2B and FIG. 2C and the known angle of illumination from light sources 104 a and 104 b can also be used to estimate the dimensional variations of element 110 .
  • it can be determined from areas in FIG. 2B and FIG. 2C in which the pixel brightness data is a maximum and does not vary that such areas are not directly blocked from direct exposure by either source.
  • an obstruction is blocking those pixels, and that the obstruction is located between light source having the lower brightness values and the location of the pixel being analyzed.
  • Other suitable procedures can be used to estimate the size and location of dimensional variations based upon pixel data, such as the use of empirically developed pass/fail ratios based upon the size of areas in which pixel brightness variations between two or more images exceed predetermined levels.
  • FIGS. 3A, 3B, and 3 C show an exemplary damaged element 110 and corresponding bright and shaded regions generated by illumination from light sources 104 a and 104 b (not explicitly shown).
  • FIG. 3A shows the damaged element 110 which varies from semi-circular, such as by an indentation. As shown in FIG. 3B, this indentation creates shaded regions 302 and 304 . These shaded regions 302 and 304 are different from shaded region 202 in FIG. 2 b . These exemplary variations can be used to detect three-dimensional variations in element 110 that would otherwise be difficult to detect from a single image, depending on the angle of illumination.
  • FIG. 3C includes shaded region 306 which varies from shaded region 204 . The pixels defining these regions can be compared between a test image, such as that shown in FIG. 3B and FIG. 3C, and a reference image, such as that shown in FIG. 2B and FIG.
  • the composite images formed by combining image data from shaded regions 202 and 204 with FIGS. 3B and 3C can be used and compared, so as to generate additional comparison points.
  • the variations in pixel brightness between the reference images and test images can also be used, in conjunction with the known angular position of light sources, to estimate the location and size of obstructions, deformations, or other features.
  • the shaded regions generated by illumination from light sources 104 a and 104 b can be used to generate three-dimensional image data from two dimensional image data.
  • Pixel brightness data can also be used to estimate the dimensional variations between a test image and a reference image.
  • FIG. 4 is a diagram of a system 400 for processing image data from multiple images in accordance with an exemplary embodiment of the present invention.
  • System 400 includes multiple image processor 102 and light sequence controller 402 , first image analyzer 404 , second image analyzer 406 , image comparator 408 , and 3D image constructor 410 , each of which can be implemented in hardware, software, or a suitable combination of hardware and software, and which can be one or more software systems operating on a general purpose processor platform.
  • Light sequence controller 402 controls the sequence in which light sources 104 a , 104 b , and other suitable lights illuminate a component 108 . Likewise, light sequence controller 402 also controls the operation of camera 106 , such that when a first light source is illuminating the component 108 , camera 106 captures first image data, and when a second light source is illuminating the component 108 , camera 106 captures or generates second image data. Likewise, light sequence controller 402 can control light sources having different frequencies, such that camera 106 can generate multiple sets of image data simultaneously so as to decrease the amount of time required to generate the multiple sets of image data.
  • First image analyzer 404 and second image analyzer 406 receive an N ⁇ M array of pixels of brightness data, and analyze the pixel data to determine whether the pixel data is acceptable, requires additional analysis such as comparison with a reference image or dimensional analysis, or is unacceptable. First image analyzer 404 and second image analyzer 406 then generate status data indicating whether the pixel data is acceptable, requires further analysis, or is unacceptable. In one exemplary embodiment, first image analyzer 404 receives pixel array data generated when light source 104 a illuminates component 108 and second image analyzer 406 receives pixel array data generated when light source 104 b illuminates component 108 . Additional image analyzers can also be used to accommodate light sources illuminating the component 108 from different angles. First image analyzer 404 and second image analyzer 406 perform pixel brightness analysis of the corresponding images.
  • first image analyzer 404 and second image analyzer 406 determine whether the pixel data indicates that the number and magnitude of variations in pixel brightness data exceed predetermined maximum allowable numbers and magnitudes, such that it is determinable whether the component contains unacceptable dimensional variations without additional image data analysis. In addition, first image analyzer 404 and second image analyzer 406 can determine whether the pixel data falls within a range of values that indicates that further analysis is required.
  • Image comparator 408 receives first image data and second image data and generates difference image data, such as by subtracting pixel brightness data for corresponding pixels between a first image and a second image.
  • Image comparator 408 can perform comparator analysis of first test image data and first reference image data, second test image data and second reference image data, composite test image data and composite reference image data, or other suitable sets of corresponding image data.
  • Image comparator 408 can also generate absolute brightness variation data, relative brightness variation data, or other suitable brightness variation data.
  • 3D image constructor 410 can receive the test image data, reference image data, difference image data, composite image data, or other suitable image data and determine whether defects, variations, or other features of element 110 or other elements exceed allowable variations for such elements.
  • 3D image constructor 410 can determine from the known angle of illumination of light sources 104 a , 104 b and other light sources, and from the brightness values of pixels generated when such light sources illuminate the component, whether the light source is illuminating the feature or element 110 at that corresponding position.
  • 3D image constructor 410 can include predetermined ranges for allowable variations, such as histogram data, pixel area mapping data, and other suitable data.
  • 3D image constructor 410 can be used to generate dimensional variation data after determining whether a variation or feature in an element 110 exceeds allowable limits, such that the component having the element can be rejected in the event the damage or dimensional variation in the element 110 exceeds such limits.
  • system 400 is used to control the inspection of a component, to generate test image data, to analyze the test image data, and to estimate three-dimensional variations or features of a test image.
  • System 400 utilizes image data generated by illuminating the component from two or more angles, can combine the test image data and compare the test image data to reference image data, and can process any difference image data to make determinations on whether or not to accept or reject a component.
  • FIG. 5 is a flowchart of a method 500 for analyzing image data from multiple images in accordance with an exemplary embodiment of the present invention.
  • Method 500 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
  • Method 500 begins at 502 where image data is obtained.
  • the image data is obtained by simultaneously illuminating a component with multiple light sources from different angles, where each light source is illuminated at a different time.
  • the light sources can provide light having different frequencies where the image data is generated at the same time and filters, tuned pixels, or other procedures are used to separate the image data created from each light source. The method then proceeds to 504 .
  • each set of image data is analyzed.
  • the sets of image data can be analyzed by generating histogram data showing the brightness of each pixel, by comparing each set of test image data to a set of reference image data and performing histogram analysis or other suitable analysis of the difference image data set, by combining the test image data and comparing the combined test image data to predetermined acceptable ranges for histogram data, by comparing the combined test image data to combined reference image data, or performing other suitable analyses.
  • the method then proceeds to 506 .
  • compared image data is obtained.
  • an initial analysis is performed at 504 to determine whether additional analyses need to be performed, such that the analysis performed at 504 does not include comparator data.
  • Other suitable processes can be used.
  • the comparator image data is obtained for each single test image and reference image, composite test and reference images, or other suitable comparative data, the method proceeds to 512 .
  • the image data and comparator data is analyzed to generate three-dimensional image data.
  • the three-dimensional image data can include predetermined allowable ranges for three-dimensional variations that generate shaded regions of elements when illuminated by multiple light sources.
  • the three-dimensional image data can include estimates of variations and components based upon the known angular relationship between the light sources and the component. The method then proceeds to 514 .
  • the three-dimensional image data is applied to template data.
  • the template data can include one or more templates that are used to estimate variations between measured brightness data and expected brightness data, so as to determine whether three-dimensional variations in the inspected component exceed allowable variations. The method then proceeds to 516 .
  • the method proceeds to 518 where the image data is rejected, such as by rejecting the component, flagging the component for further manual inspection, or other suitable procedures. Otherwise, the method proceeds to 520 where the image data is accepted.
  • method 500 is used to analyze multiple sets of image data for a test component in order to determine whether the component includes dimensional variations, damage, or other unacceptable condition.
  • Method 500 further utilizes light sources having different angular relationships to the test component, where the known angular relationship of the light sources can be used in conjunction with the pixel brightness data to estimate 3-dimensional variations in the test component.
  • FIG. 6 is a flowchart of a method 600 for analyzing image data in accordance with an exemplary embodiment of the present invention.
  • Method 600 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
  • Method 600 begins at 602 where a test piece is exposed to light from two different light frequencies and two different angular illumination zones. The method then proceeds to 604 and 608 in parallel.
  • first image data is obtained, such as by filtering the light through a first filter, by using pixels tuned to the first light frequency, or other suitable methods.
  • second image data is obtained, such as by filtering the light through a second filter, by using pixels tuned to a second light frequency, or other suitable methods.
  • the method then proceeds to 606 from 604 and 610 from 608 , respectively.
  • the pixel brightness variation data is analyzed for the first image data. For example, pixel histogram data can be generated and the variations in pixel brightness can be compared to predetermined acceptable ranges. Likewise, other suitable pixel brightness variation analysis methods can be used.
  • similar pixel brightness variations are analyzed for the second image data. The method then proceeds to 612 and 614 , respectively.
  • method 600 can be used to determine whether three-dimensional analysis of component image data should be performed, such as to perform a quick preliminary component image inspection analysis for the purpose of determining whether additional analysis should be performed.
  • Method 600 also allows image data to be analyzed in parallel where suitable, such as when a parallel processor platform is being used to analyze the image data.
  • method 600 can be implemented on a general purpose processing platform, such as through the use of multitasking or by otherwise simulating parallel processing.
  • FIG. 7 is a flowchart of a method 700 for performing image data analysis for multiple images in accordance with an exemplary embodiment of the present invention.
  • Method 700 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
  • Method 700 begins at 702 and 704 in parallel.
  • first reference image data is compared to first test image data
  • second reference image data is compared to second test image data. This comparison can include a pixel to correlating pixel brightness subtraction to generate a difference image, or other suitable comparison procedures.
  • the method then proceeds to 706 .
  • a composite test image is formed.
  • the composite test image can include two or more sets of image data generated from two or more different illumination angles, from two or more different light frequencies, or other suitable composite test data. The method then proceeds to 712 .
  • the composite test image data is compared to composite reference image data, such as by performing a pixel to corresponding pixel subtraction or other suitable compare procedures. The method then proceeds to 714 .
  • three-dimensional coordinates for the component being inspected are estimated from variations in the test image data as compared to the reference image data. For example, pixels at coordinates that have significant variations in brightness as a function of the angle of illumination can indicate the existence of an indentation, spur, bulge, or other deformity in the component. It may be determined by analysis, empirically, or otherwise that such variations in brightness that exceed certain levels correlate to dimensional variations. Likewise, an estimate of the dimensional variation can be calculated from the brightness data and the known angular position of each light source. Other suitable methods may also or alternatively be used. The method then proceeds to 716 .
  • allowable variations can be determined empirically, by calculation, can be set according to a customer or industry standard, or through other suitable methods. The method then proceeds to 718 if it is determined that the variations exceed allowable ranges and the image data is rejected. Likewise, a message can be generated informing the operator that additional analysis or operator inspection is required. If it is determined at 716 that any variations in the image data are allowable then the method proceeds to 720 where the image data is accepted.
  • method 700 allows a component to be inspected by illuminating the component from multiple light sources, such that the component generates shaded regions and bright regions.
  • the shaded and bright regions of the component can be then analyzed and compared to reference image data to determine whether unacceptable variations or damage may exist on the component.

Abstract

A system for analyzing multiple images is provided, such as to locate defects in a test component. The system includes a first light source, such as one that emits blue light, and a second light source, such as one that emits red light. The system also includes a camera, where the camera and the light sources are focused on an area where a test piece is to be placed. A multiple image processor is connected to the first light source, the second light source, and the camera. The multiple image processor causes the first light source and the second light source to turn on, such as in sequence, and also causes the camera to generate two or more sets of image data, such as one set when each of the light sources is illuminated, through the use of filters or tuned pixels, or otherwise.

Description

    FIELD OF THE INVENTION
  • The present invention pertains to the field of semiconductor devices, and more particularly to a system and method for inspecting semiconductor devices that uses multiple two-dimensional images to generate third dimension data. [0001]
  • BACKGROUND OF THE INVENTION
  • Image data analysis systems for inspecting semiconductor components are known in the art. Such image data analysis systems attempt to determine the state of the semiconducting component or other inspected components by analyzing image data, which is typically comprised of an N×M array of picture elements or “pixels.” The brightness value of each pixel of a test image is typically compared to the brightness element of a corresponding pixel of a reference image, and the comparison data is analyzed to determine whether or not unacceptable defects exist on the semiconducting device, component or other object being inspected. For example, image data analysis is used to determine whether the variation in the dimensions of an element of the component exceed allowable tolerances for such dimensions. [0002]
  • One drawback with known image data inspection systems is the difficulty in determining the three-dimensional nature of elements. Such image data is typically taken from a single angle, such that any three-dimensional aspects or flaws may be difficult to detect. For example, a common method for determining the three-dimensional aspects of a semiconductor device or component that is being inspected is to use a laser beam to trace a line, and to determine when the line varies from a straight line, where such variations are then correlated to defects in the semiconducting device or component. When the semiconducting device or component contains a large number of elements, it is necessary to trace a laser line through each of the elements, which can require movement of the component to a number of different locations. Likewise, it is possible that the laser drawn line may not lie on a defect, such that the defect could be missed. [0003]
  • Thus, although it is known to perform analysis of image data of a component to determine whether variations in the dimensions of elements of the component exceed allowable tolerances, the determination of such dimensional variations in three dimensions is time-consuming and limited to small portions of the component. [0004]
  • BRIEF SUMMARY OF THE INVENTION
  • In accordance with the present invention, a system and method for multiple image analysis are provided that overcome known problems with analyzing image data. [0005]
  • In particular, a system and method for multiple image analysis are provided that use image data generated by illuminating a component from two or more lighting angles, which allows three-dimensional aspects of the component to be determined. [0006]
  • In accordance with an exemplary embodiment of the present invention, a system for analyzing multiple images is provided, such as to locate defects in a test component. The system includes a first light source, such as one that emits blue light, and a second light source, such as one that emits red light. The system also includes a camera, where the camera and the light sources are focused on an area where a test piece is to be placed. A multiple image processor is connected to the first light source, the second light source, and the camera. The multiple image processor causes the first light source and the second light source to turn on, such as in sequence, and also causes the camera to generate two or more sets of image data, such as one set when each of the light sources is illuminated, through the use of filters or tuned pixels, or otherwise. [0007]
  • The present invention provides many important technical advantages. One important technical advantage of the present invention is a system and method for multiple image analysis that uses two or more sets of image data to analyze a component. Each set of image data is obtained when the component is illuminated by a light source having a different lighting angle, which creates shaded areas that can be analyzed to determine whether they indicate the existence of damage or unacceptable dimensional variations on the component. [0008]
  • Those skilled in the art will further appreciate the advantages and superior features of the invention together with other important aspects thereof on reading the detailed description that follows in conjunction with the drawings. [0009]
  • BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
  • FIG. 1 is a diagram of a system for performing multiple image analysis in accordance with an exemplary embodiment of the present invention; [0010]
  • FIGS. 2A, 2B, and [0011] 2C show an exemplary undamaged element and corresponding bright and shaded regions generated by illumination from light sources;
  • FIGS. 3A, 3B, and [0012] 3C show an exemplary damaged element and corresponding bright and shaded regions generated by illumination from light sources;
  • FIG. 4 is a diagram of a system for processing image data from multiple images in accordance with an exemplary embodiment of the present invention; [0013]
  • FIG. 5 is a flowchart of a method for analyzing image data from multiple images in accordance with an exemplary embodiment of the present invention; [0014]
  • FIG. 6 is a flowchart of a method for analyzing image data in accordance with an exemplary embodiment of the present invention; and [0015]
  • FIG. 7 is a flowchart of a method for performing image data analysis for multiple images in accordance with an exemplary embodiment of the present invention. [0016]
  • DETAILED DESCRIPTION OF THE INVENTION
  • In the description that follows, like parts are marked throughout the specification and drawings with the same reference numerals, respectively. The drawing figures might not be to scale, and certain components can be shown in generalized or schematic form and identified by commercial designations in the interest of clarity and conciseness. [0017]
  • FIG. 1 is a diagram of a system [0018] 100 for performing multiple image analysis in accordance with an exemplary embodiment of the present invention. System 100 allows three-dimensional aspects of an inspected device or component to be determined from images obtained from two or more different viewing angles.
  • System [0019] 100 includes multiple image processor 102, which can be implemented in hardware, software, or a suitable combination of hardware and software, and which can be one or more software systems operating on a general purpose processing platform. As used herein, a software system can include one or more objects, agents, subroutines, lines of code, threads, two or more lines of code or other suitable software structures operating in two or more separate software applications, or other suitable software structures, and can operate on two or more different processors, or other suitable configurations of processors. In one exemplary embodiment, a software system can include one or more lines of code or other software structures operating in a general purpose software application, such as an operating system, and one or more lines of code or other suitable software structures operating in a specific purpose software application.
  • [0020] Multiple image processor 102 is coupled to light sources 104 a and 104 b. As used herein, the term “couple,” and its cognate terms such as “couples” and “coupled,” can include a physical connection (such as a copper conductor), a virtual connection (such as through one or more randomly assigned data memory locations of a data memory device), a logical connection (such as through one or more logical gates of a semiconducting device), a wireless connection, other suitable connections, or a suitable combination of such connections. In one exemplary embodiment, systems and components are coupled to other systems and components through intervening systems and components, such as through an operating system of a general purpose processor platform.
  • [0021] Multiple image processor 102 is also coupled to camera 106. Camera 106 can be a charge coupled device (CCD), a CMOS imaging device, or other suitable imaging devices that are focused on a component 108 having a plurality of elements 110. Light sources 104 a and 104 b are also focused on component 108, and illuminate component 108 from different angles as shown in FIG. 1. Thus, the light illuminating component 108 from light source 104 a will create shaded regions that are different from the shaded regions created by light illuminating component 108 from light source 104 b. Additional light sources can be used where suitable to create additional shaded regions.
  • Camera [0022] 106 is used to record image data of component 108 while it is being illuminated by light sources 104 a and 104 b. In one exemplary embodiment, camera 106 is controlled by multiple image processor 102 to store a first set of image data of component 108 when light source 104 a is on, and to store a second set of image data when light source 104 b is on. Likewise, camera 106 can store image data when both of light sources 104 a and 104 b are on, such as when the light sources use different frequencies of light. For example, camera 106 can record image data according to the frequency of the light that creates the image, such as by including one or more light filters, two or more sets of pixels that are tuned to received predetermined frequencies of light, or to otherwise differentiate between light illuminated from light sources 104 a and 104 b, such that multiple sets of image data can be concurrently gathered.
  • In operation, a [0023] component 108 is placed in the focal area of light sources 104 a and 104 b and camera 106 for inspection. Multiple image processor 102 then causes component 108 to be illuminated and causes camera 106 to produce image data, such as by generating an N×M array of pixels of image data, which can then be stored by multiple image processor 102 or other suitable storage systems or devices. Because of the angular difference between light sources 104 a and 104 b relative to component 108, shaded regions are generated from elements 110. Multiple image processor 102 can analyze these shaded regions to determine whether they are indicative of any damage or defects to component 108, elements 110, or other suitable indications.
  • In this manner, [0024] multiple image processor 102 can determine whether three-dimensional defects or other variations in component 108 or elements 110 exist. For example, if one of elements 110 is damaged, then the shaded regions generated by that element 110 when it is illuminated by light sources 104 a and 104 b will vary from the shaded regions generated for undamaged reference images. Furthermore, the variations in pixel brightness between corresponding pixels of the test image data and the reference image data, as illuminated by light sources 104 a and 104 b, can also be analyzed to generate an approximation of differences in height, dimensions, or other data that can be used to approximate a three-dimensional analysis.
  • FIGS. 2A, 2B, and [0025] 2C show an exemplary undamaged element 110 and corresponding bright and shaded regions generated by illumination from light sources 104 a and 104 b (not explicitly shown).
  • FIG. 2A shows an exemplary [0026] undamaged element 110, which is semi-spherical in shape. In FIG. 2B, the circular outline of element 110 as viewed from overhead is shown with an illuminated region and a shaded region corresponding to the shadow generated by light source 104 a. As shown, the shaded region generates a distinctive pattern which is indicative of a spherical configuration of element 110. Likewise, in FIG. 2C, the shaded region of element 110 is on the opposite face, as a result of the location of light source 104 b. Thus, the shaded regions generated shown in FIGS. 2b and 2 c can be used as a reference for an undamaged element 110.
  • In addition, the differences in pixel brightness data between FIG. 2B and FIG. 2C and the known angle of illumination from [0027] light sources 104 a and 104 b can also be used to estimate the dimensional variations of element 110. For example, it can be determined from areas in FIG. 2B and FIG. 2C in which the pixel brightness data is a maximum and does not vary that such areas are not directly blocked from direct exposure by either source. Likewise, as the difference in brightness data increases for a given pixel of FIG. 2B and FIG. 2C, it can be determined that an obstruction is blocking those pixels, and that the obstruction is located between light source having the lower brightness values and the location of the pixel being analyzed. Other suitable procedures can be used to estimate the size and location of dimensional variations based upon pixel data, such as the use of empirically developed pass/fail ratios based upon the size of areas in which pixel brightness variations between two or more images exceed predetermined levels.
  • FIGS. 3A, 3B, and [0028] 3C show an exemplary damaged element 110 and corresponding bright and shaded regions generated by illumination from light sources 104 a and 104 b (not explicitly shown).
  • FIG. 3A shows the damaged [0029] element 110 which varies from semi-circular, such as by an indentation. As shown in FIG. 3B, this indentation creates shaded regions 302 and 304. These shaded regions 302 and 304 are different from shaded region 202 in FIG. 2b. These exemplary variations can be used to detect three-dimensional variations in element 110 that would otherwise be difficult to detect from a single image, depending on the angle of illumination. Likewise, FIG. 3C includes shaded region 306 which varies from shaded region 204. The pixels defining these regions can be compared between a test image, such as that shown in FIG. 3B and FIG. 3C, and a reference image, such as that shown in FIG. 2B and FIG. 2C, to determine whether the region defined by such pixel variations exceeds predetermined allowable areas for defects. Likewise, the composite images formed by combining image data from shaded regions 202 and 204 with FIGS. 3B and 3C can be used and compared, so as to generate additional comparison points. The variations in pixel brightness between the reference images and test images can also be used, in conjunction with the known angular position of light sources, to estimate the location and size of obstructions, deformations, or other features.
  • In operation, the shaded regions generated by illumination from [0030] light sources 104 a and 104 b can be used to generate three-dimensional image data from two dimensional image data. Pixel brightness data can also be used to estimate the dimensional variations between a test image and a reference image.
  • FIG. 4 is a diagram of a system [0031] 400 for processing image data from multiple images in accordance with an exemplary embodiment of the present invention. System 400 includes multiple image processor 102 and light sequence controller 402, first image analyzer 404, second image analyzer 406, image comparator 408, and 3D image constructor 410, each of which can be implemented in hardware, software, or a suitable combination of hardware and software, and which can be one or more software systems operating on a general purpose processor platform.
  • [0032] Light sequence controller 402 controls the sequence in which light sources 104 a, 104 b, and other suitable lights illuminate a component 108. Likewise, light sequence controller 402 also controls the operation of camera 106, such that when a first light source is illuminating the component 108, camera 106 captures first image data, and when a second light source is illuminating the component 108, camera 106 captures or generates second image data. Likewise, light sequence controller 402 can control light sources having different frequencies, such that camera 106 can generate multiple sets of image data simultaneously so as to decrease the amount of time required to generate the multiple sets of image data.
  • [0033] First image analyzer 404 and second image analyzer 406 receive an N×M array of pixels of brightness data, and analyze the pixel data to determine whether the pixel data is acceptable, requires additional analysis such as comparison with a reference image or dimensional analysis, or is unacceptable. First image analyzer 404 and second image analyzer 406 then generate status data indicating whether the pixel data is acceptable, requires further analysis, or is unacceptable. In one exemplary embodiment, first image analyzer 404 receives pixel array data generated when light source 104 a illuminates component 108 and second image analyzer 406 receives pixel array data generated when light source 104 b illuminates component 108. Additional image analyzers can also be used to accommodate light sources illuminating the component 108 from different angles. First image analyzer 404 and second image analyzer 406 perform pixel brightness analysis of the corresponding images.
  • In one exemplary embodiment, [0034] first image analyzer 404 and second image analyzer 406 determine whether the pixel data indicates that the number and magnitude of variations in pixel brightness data exceed predetermined maximum allowable numbers and magnitudes, such that it is determinable whether the component contains unacceptable dimensional variations without additional image data analysis. In addition, first image analyzer 404 and second image analyzer 406 can determine whether the pixel data falls within a range of values that indicates that further analysis is required.
  • [0035] Image comparator 408 receives first image data and second image data and generates difference image data, such as by subtracting pixel brightness data for corresponding pixels between a first image and a second image. Image comparator 408 can perform comparator analysis of first test image data and first reference image data, second test image data and second reference image data, composite test image data and composite reference image data, or other suitable sets of corresponding image data. Image comparator 408 can also generate absolute brightness variation data, relative brightness variation data, or other suitable brightness variation data.
  • [0036] 3D image constructor 410 can receive the test image data, reference image data, difference image data, composite image data, or other suitable image data and determine whether defects, variations, or other features of element 110 or other elements exceed allowable variations for such elements. In one exemplary embodiment, 3D image constructor 410 can determine from the known angle of illumination of light sources 104 a, 104 b and other light sources, and from the brightness values of pixels generated when such light sources illuminate the component, whether the light source is illuminating the feature or element 110 at that corresponding position. Likewise, 3D image constructor 410 can include predetermined ranges for allowable variations, such as histogram data, pixel area mapping data, and other suitable data. In this manner, 3D image constructor 410 can be used to generate dimensional variation data after determining whether a variation or feature in an element 110 exceeds allowable limits, such that the component having the element can be rejected in the event the damage or dimensional variation in the element 110 exceeds such limits.
  • In operation, system [0037] 400 is used to control the inspection of a component, to generate test image data, to analyze the test image data, and to estimate three-dimensional variations or features of a test image. System 400 utilizes image data generated by illuminating the component from two or more angles, can combine the test image data and compare the test image data to reference image data, and can process any difference image data to make determinations on whether or not to accept or reject a component.
  • FIG. 5 is a flowchart of a method [0038] 500 for analyzing image data from multiple images in accordance with an exemplary embodiment of the present invention. Method 500 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
  • Method [0039] 500 begins at 502 where image data is obtained. In one exemplary embodiment, the image data is obtained by simultaneously illuminating a component with multiple light sources from different angles, where each light source is illuminated at a different time. Likewise, the light sources can provide light having different frequencies where the image data is generated at the same time and filters, tuned pixels, or other procedures are used to separate the image data created from each light source. The method then proceeds to 504.
  • At [0040] 504 each set of image data is analyzed. The sets of image data can be analyzed by generating histogram data showing the brightness of each pixel, by comparing each set of test image data to a set of reference image data and performing histogram analysis or other suitable analysis of the difference image data set, by combining the test image data and comparing the combined test image data to predetermined acceptable ranges for histogram data, by comparing the combined test image data to combined reference image data, or performing other suitable analyses. The method then proceeds to 506.
  • At [0041] 506 it is determined whether all images are within an acceptable predetermined range. If all images are within such range, the method proceeds to 508 where the image is accepted and any subsequent analysis is performed on that component, other components can be selected for analysis, or other suitable procedures can be implemented. Otherwise the method proceeds to 510.
  • At [0042] 510 compared image data is obtained. In one exemplary embodiment, an initial analysis is performed at 504 to determine whether additional analyses need to be performed, such that the analysis performed at 504 does not include comparator data. Other suitable processes can be used. After the comparator image data is obtained for each single test image and reference image, composite test and reference images, or other suitable comparative data, the method proceeds to 512.
  • At [0043] 512 the image data and comparator data is analyzed to generate three-dimensional image data. In one exemplary embodiment, the three-dimensional image data can include predetermined allowable ranges for three-dimensional variations that generate shaded regions of elements when illuminated by multiple light sources. Likewise, the three-dimensional image data can include estimates of variations and components based upon the known angular relationship between the light sources and the component. The method then proceeds to 514.
  • At [0044] 514 the three-dimensional image data is applied to template data. In one exemplary embodiment, the template data can include one or more templates that are used to estimate variations between measured brightness data and expected brightness data, so as to determine whether three-dimensional variations in the inspected component exceed allowable variations. The method then proceeds to 516.
  • At [0045] 516 it is determined whether the three-dimensional data is within a predetermined range. If the three-dimensional data is not within the range, the method proceeds to 518 where the image data is rejected, such as by rejecting the component, flagging the component for further manual inspection, or other suitable procedures. Otherwise, the method proceeds to 520 where the image data is accepted.
  • In operation, method [0046] 500 is used to analyze multiple sets of image data for a test component in order to determine whether the component includes dimensional variations, damage, or other unacceptable condition. Method 500 further utilizes light sources having different angular relationships to the test component, where the known angular relationship of the light sources can be used in conjunction with the pixel brightness data to estimate 3-dimensional variations in the test component.
  • FIG. 6 is a flowchart of a method [0047] 600 for analyzing image data in accordance with an exemplary embodiment of the present invention. Method 600 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
  • Method [0048] 600 begins at 602 where a test piece is exposed to light from two different light frequencies and two different angular illumination zones. The method then proceeds to 604 and 608 in parallel. At 604, first image data is obtained, such as by filtering the light through a first filter, by using pixels tuned to the first light frequency, or other suitable methods. Likewise, at 608, the second image data is obtained, such as by filtering the light through a second filter, by using pixels tuned to a second light frequency, or other suitable methods. The method then proceeds to 606 from 604 and 610 from 608, respectively.
  • At [0049] 606, the pixel brightness variation data is analyzed for the first image data. For example, pixel histogram data can be generated and the variations in pixel brightness can be compared to predetermined acceptable ranges. Likewise, other suitable pixel brightness variation analysis methods can be used. At 610, similar pixel brightness variations are analyzed for the second image data. The method then proceeds to 612 and 614, respectively.
  • At [0050] 612 and 614 it is determined whether the variations in pixel brightness for the first image data and the second image data are within a predetermined range. If both sets of image data have acceptable variations then the method proceeds to 618 and the image data is accepted. Likewise, if either of the images has unacceptable range variations the method proceeds to 616 where three-dimensional analysis is performed to determine whether the dimensional variations in the test piece are acceptable.
  • In operation, method [0051] 600 can be used to determine whether three-dimensional analysis of component image data should be performed, such as to perform a quick preliminary component image inspection analysis for the purpose of determining whether additional analysis should be performed. Method 600 also allows image data to be analyzed in parallel where suitable, such as when a parallel processor platform is being used to analyze the image data. Alternatively, method 600 can be implemented on a general purpose processing platform, such as through the use of multitasking or by otherwise simulating parallel processing.
  • FIG. 7 is a flowchart of a [0052] method 700 for performing image data analysis for multiple images in accordance with an exemplary embodiment of the present invention. Method 700 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
  • [0053] Method 700 begins at 702 and 704 in parallel. At 702, first reference image data is compared to first test image data, and at 704, second reference image data is compared to second test image data. This comparison can include a pixel to correlating pixel brightness subtraction to generate a difference image, or other suitable comparison procedures. The method then proceeds to 706.
  • At [0054] 706 it is determined whether acceptable variations exist in the compare data, such as by generating a histogram having pixel frequency and magnitude for difference data. If it is determined that the variations are acceptable the method proceeds to 708 where the image data is accepted. Likewise, if the variations are not acceptable the method proceeds to 710.
  • At [0055] 710 a composite test image is formed. In one exemplary embodiment, the composite test image can include two or more sets of image data generated from two or more different illumination angles, from two or more different light frequencies, or other suitable composite test data. The method then proceeds to 712.
  • At [0056] 712 the composite test image data is compared to composite reference image data, such as by performing a pixel to corresponding pixel subtraction or other suitable compare procedures. The method then proceeds to 714.
  • At [0057] 714, three-dimensional coordinates for the component being inspected are estimated from variations in the test image data as compared to the reference image data. For example, pixels at coordinates that have significant variations in brightness as a function of the angle of illumination can indicate the existence of an indentation, spur, bulge, or other deformity in the component. It may be determined by analysis, empirically, or otherwise that such variations in brightness that exceed certain levels correlate to dimensional variations. Likewise, an estimate of the dimensional variation can be calculated from the brightness data and the known angular position of each light source. Other suitable methods may also or alternatively be used. The method then proceeds to 716.
  • At [0058] 716, it is determined whether the variations in the component dimensions are allowable. In one exemplary embodiment, allowable variations can be determined empirically, by calculation, can be set according to a customer or industry standard, or through other suitable methods. The method then proceeds to 718 if it is determined that the variations exceed allowable ranges and the image data is rejected. Likewise, a message can be generated informing the operator that additional analysis or operator inspection is required. If it is determined at 716 that any variations in the image data are allowable then the method proceeds to 720 where the image data is accepted.
  • In operation, [0059] method 700 allows a component to be inspected by illuminating the component from multiple light sources, such that the component generates shaded regions and bright regions. The shaded and bright regions of the component can be then analyzed and compared to reference image data to determine whether unacceptable variations or damage may exist on the component.
  • Although exemplary embodiments of a system and method for multiple image analysis have been described in detail herein, those skilled in the art will also recognize that various substitutions and modifications can be made to the systems and methods without departing from the scope and spirit of the appended claims. [0060]

Claims (20)

What is claimed is
1. A system for multiple image analysis comprising:
a first light source;
a second light source;
a camera; and
a multiple image processor coupled to the first light source, the second light source, and the camera, the multiple image processor causing the first light source and the second light source to turn on and the camera to generate two or more sets of image data.
2. The system of claim 1 wherein the first light source emits light having a first frequency and the second light source emits light having a second frequency.
3. The system of claim 2 wherein the camera can generate two or more sets of image data when both the first light source and the second light source are emitting light.
4. The system of claim 2 wherein the camera further comprises:
a first set of pixels receiving light at the first frequency; and
a second set of pixels receiving light at the second frequency.
5. The system of claim 2 wherein the camera further comprises:
a first filter passing light at the first frequency; and
a second filter passing light at the second frequency.
6. The system of claim 1 wherein the multiple image processor further comprises a light sequence controller causing the first light source and the second light source to turn on and turn off.
7. The system of claim 1 wherein the multiple image processor further comprises an image analyzer receiving the two or more sets of image data and generating status data that indicates whether the image data is acceptable.
8. The system of claim 1 wherein the multiple image processor further comprises a first image analyzer receiving the first set of image data and a second image analyzer receiving the second set of image data and generating status data that indicates whether the image data is acceptable.
9. The system of claim 1 wherein the multiple image processor further comprises an image comparator receiving the two or more sets of image data and generating difference data.
10. The system of claim 1 wherein the multiple image processor further comprises an image constructor receiving the two or more sets of image data and generating dimensional variation data.
11. A method for inspecting a component comprising:
illuminating the component from a first illumination angle;
receiving first image data of the component;
illuminating the component from a second illumination angle;
receiving second image data of the component; and
using the first image data and the second image data to determine whether a dimension of the component is acceptable.
12. The method of claim 11 wherein illuminating the component from the first illumination angle and illuminating the component from the second illumination angle further comprises illuminating the component using light having a first frequency from the first illumination angle and illuminating the component using light having a second frequency from the second illumination angle.
13. The method of claim 11 wherein receiving the first image data of the component comprises receiving the first image data of the component by filtering light received from the component.
14. The method of claim 11 wherein receiving the first image data of the component and receiving the second image data of the component comprises receiving the first image data of the component by filtering light received from the component with a first filter and receiving the second image data of the component by filtering light received from the component with a second filter.
15. The method of claim 11 wherein receiving the first image data of the component comprises receiving the first image data of the component with a first set of pixels.
16. The method of claim 11 wherein receiving the first image data of the component and receiving the second image data of the component comprises receiving the first image data of the component with a first set of pixels and receiving the second image data of the component with a second set of pixels.
17. A method for inspecting a component comprising:
receiving first image data and second image data of the component;
comparing the first image data to reference image data to generate first difference data;
comparing the second image data to reference image data to generate second difference data; and
generating component dimension data from the first difference data and the second difference data.
18. The method of claim 17 further comprising:
combining the first image data and the second image data to generate composite image data;
comparing the composite image data to composite reference data to generate composite difference data; and
generating component dimension data from the composite difference data.
19. The method of claim 17 wherein the step of receiving the first image data and the second image data of the component is preceded by the step of receiving status data that indicates that the component requires additional analysis to determine whether it has unacceptable dimensional variations.
20. The method of claim 17 wherein generating the component dimension data from the first difference data and the second difference data further comprises using light source angular data to generate the component dimension data.
US09/876,795 2001-06-07 2001-06-07 System and method for multiple image analysis Abandoned US20020186878A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US09/876,795 US20020186878A1 (en) 2001-06-07 2001-06-07 System and method for multiple image analysis
PCT/IB2002/002050 WO2002099403A1 (en) 2001-06-07 2002-06-05 System and method for multiple image analysis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US09/876,795 US20020186878A1 (en) 2001-06-07 2001-06-07 System and method for multiple image analysis

Publications (1)

Publication Number Publication Date
US20020186878A1 true US20020186878A1 (en) 2002-12-12

Family

ID=25368601

Family Applications (1)

Application Number Title Priority Date Filing Date
US09/876,795 Abandoned US20020186878A1 (en) 2001-06-07 2001-06-07 System and method for multiple image analysis

Country Status (2)

Country Link
US (1) US20020186878A1 (en)
WO (1) WO2002099403A1 (en)

Cited By (81)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030169917A1 (en) * 2002-01-15 2003-09-11 Nec Corporation Three-dimensional body photography apparatus, three-dimensional geometricmodel generation apparatus, three-dimensional geometric model generation method, and record medium in which three-dimensional geometric model generation program was stored
US20040168881A1 (en) * 2003-01-10 2004-09-02 National Rejectors Inc. Gmbh Process for identifying an embossed image of a coin in an automatice coin machine
US20040245334A1 (en) * 2003-06-06 2004-12-09 Sikorski Steven Maurice Inverted terminal presentation scanner and holder
US20050103753A1 (en) * 2003-11-18 2005-05-19 Matsushita Electric Industrial Co., Ltd. Light processing apparatus
US20050286754A1 (en) * 2004-06-14 2005-12-29 Thomas Steven P Electronic component detection system
US20060006311A1 (en) * 2001-06-22 2006-01-12 Orbotech Ltd High-sensitivity optical scanning using memory integration
US20060291714A1 (en) * 2004-12-07 2006-12-28 Kenong Wu Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle
GB2427541A (en) * 2005-06-24 2007-01-03 Aew Delford Group Ltd Vision system for food cutting and portioning apparatus
GB2428182A (en) * 2005-06-24 2007-01-24 Aew Delford Systems Vision system for food cutting and portioning apparatus
US20070027585A1 (en) * 2005-07-27 2007-02-01 Thomas Wulff System and method for monitoring a mobile computing product/arrangement
US20070076943A1 (en) * 2005-04-15 2007-04-05 Vistec Semiconductor Systems Jena Gmbh Method and apparatus for inspecting a wafer
US20070297028A1 (en) * 2006-06-21 2007-12-27 Thomas Wulff System and device for monitoring a computing device
US20070298751A1 (en) * 2006-06-21 2007-12-27 Thomas Wulff System and method for monitoring a mobile device
US7646906B2 (en) 2004-01-29 2010-01-12 Kla-Tencor Technologies Corp. Computer-implemented methods for detecting defects in reticle design data
US7676077B2 (en) 2005-11-18 2010-03-09 Kla-Tencor Technologies Corp. Methods and systems for utilizing design data in combination with inspection data
US7689966B2 (en) 2004-09-14 2010-03-30 Kla-Tencor Technologies Corp. Methods, systems, and carrier media for evaluating reticle layout data
US20100100623A1 (en) * 2004-04-06 2010-04-22 Thomas Wulff System and method for monitoring a mobile computing product/arrangement
US7711514B2 (en) 2007-08-10 2010-05-04 Kla-Tencor Technologies Corp. Computer-implemented methods, carrier media, and systems for generating a metrology sampling plan
US7738093B2 (en) 2007-05-07 2010-06-15 Kla-Tencor Corp. Methods for detecting and classifying defects on a reticle
US7769225B2 (en) * 2005-08-02 2010-08-03 Kla-Tencor Technologies Corp. Methods and systems for detecting defects in a reticle design pattern
US7796804B2 (en) 2007-07-20 2010-09-14 Kla-Tencor Corp. Methods for generating a standard reference die for use in a die to standard reference die inspection and methods for inspecting a wafer
US20100245813A1 (en) * 2009-03-24 2010-09-30 Orbotech Ltd. Multi-modal imaging
US7877722B2 (en) 2006-12-19 2011-01-25 Kla-Tencor Corp. Systems and methods for creating inspection recipes
US7962863B2 (en) 2007-05-07 2011-06-14 Kla-Tencor Corp. Computer-implemented methods, systems, and computer-readable media for determining a model for predicting printability of reticle features on a wafer
US7975245B2 (en) 2007-08-20 2011-07-05 Kla-Tencor Corp. Computer-implemented methods for determining if actual defects are potentially systematic defects or potentially random defects
US20110206237A1 (en) * 2010-02-25 2011-08-25 Canon Kabushiki Kaisha Recognition apparatus and method thereof, and computer program
US8041103B2 (en) 2005-11-18 2011-10-18 Kla-Tencor Technologies Corp. Methods and systems for determining a position of inspection data in design data space
US8112241B2 (en) 2009-03-13 2012-02-07 Kla-Tencor Corp. Methods and systems for generating an inspection process for a wafer
US20120031975A1 (en) * 2010-08-04 2012-02-09 The Code Corporation Illumination blocks for a graphical code reader
US8139844B2 (en) 2008-04-14 2012-03-20 Kla-Tencor Corp. Methods and systems for determining a defect criticality index for defects on wafers
US8194968B2 (en) 2007-01-05 2012-06-05 Kla-Tencor Corp. Methods and systems for using electrical information for a device being fabricated on a wafer to perform one or more defect-related functions
US8204297B1 (en) 2009-02-27 2012-06-19 Kla-Tencor Corp. Methods and systems for classifying defects detected on a reticle
US8213704B2 (en) 2007-05-09 2012-07-03 Kla-Tencor Corp. Methods and systems for detecting defects in a reticle design pattern
US20130144797A1 (en) * 2008-10-02 2013-06-06 ecoATM, Inc. Method And Apparatus For Recycling Electronic Devices
CN103674974A (en) * 2012-09-06 2014-03-26 株式会社岛津制作所 Inspection apparatus
US8775101B2 (en) 2009-02-13 2014-07-08 Kla-Tencor Corp. Detecting defects on a wafer
US8781781B2 (en) 2010-07-30 2014-07-15 Kla-Tencor Corp. Dynamic care areas
US8826200B2 (en) 2012-05-25 2014-09-02 Kla-Tencor Corp. Alteration for wafer inspection
US8831334B2 (en) 2012-01-20 2014-09-09 Kla-Tencor Corp. Segmentation for wafer inspection
US8923600B2 (en) 2005-11-18 2014-12-30 Kla-Tencor Technologies Corp. Methods and systems for utilizing design data in combination with inspection data
US9053527B2 (en) 2013-01-02 2015-06-09 Kla-Tencor Corp. Detecting defects on a wafer
US9087367B2 (en) 2011-09-13 2015-07-21 Kla-Tencor Corp. Determining design coordinates for wafer defects
US9092846B2 (en) 2013-02-01 2015-07-28 Kla-Tencor Corp. Detecting defects on a wafer using defect-specific and multi-channel information
US9134254B2 (en) 2013-01-07 2015-09-15 Kla-Tencor Corp. Determining a position of inspection system output in design data space
US9170211B2 (en) 2011-03-25 2015-10-27 Kla-Tencor Corp. Design-based inspection using repeating structures
US9189844B2 (en) 2012-10-15 2015-11-17 Kla-Tencor Corp. Detecting defects on a wafer using defect-specific information
JP2016035405A (en) * 2014-08-01 2016-03-17 リコーエレメックス株式会社 Image inspection device, image inspection system, and image inspection method
US9311698B2 (en) 2013-01-09 2016-04-12 Kla-Tencor Corp. Detecting defects on a wafer using template image matching
US9310320B2 (en) 2013-04-15 2016-04-12 Kla-Tencor Corp. Based sampling and binning for yield critical defects
US20160267339A1 (en) * 2015-03-11 2016-09-15 Canon Kabushiki Kaisha Image processing apparatus and method of generating face image
JP2017032548A (en) * 2015-06-26 2017-02-09 コグネックス・コーポレイション Using 3d vision for automated industrial inspection
US9659670B2 (en) 2008-07-28 2017-05-23 Kla-Tencor Corp. Computer-implemented methods, computer-readable media, and systems for classifying defects detected in a memory device area on a wafer
JP2017122682A (en) * 2016-01-08 2017-07-13 株式会社キーエンス Visual inspection device, visual inspection method, and computer program executable by controller used in visual inspection device
US9818160B2 (en) 2008-10-02 2017-11-14 ecoATM, Inc. Kiosk for recycling electronic devices
US9865512B2 (en) 2013-04-08 2018-01-09 Kla-Tencor Corp. Dynamic design attributes for wafer inspection
US9881284B2 (en) 2008-10-02 2018-01-30 ecoATM, Inc. Mini-kiosk for recycling electronic devices
US9885672B2 (en) 2016-06-08 2018-02-06 ecoATM, Inc. Methods and systems for detecting screen covers on electronic devices
US9904911B2 (en) 2008-10-02 2018-02-27 ecoATM, Inc. Secondary market and vending system for devices
US9911102B2 (en) 2014-10-02 2018-03-06 ecoATM, Inc. Application for device evaluation and other processes associated with device recycling
TWI626438B (en) * 2015-09-30 2018-06-11 佳能股份有限公司 Inspection apparatus and article manufacturing method
US10032140B2 (en) * 2008-10-02 2018-07-24 ecoATM, LLC. Systems for recycling consumer electronic devices
US10127647B2 (en) 2016-04-15 2018-11-13 Ecoatm, Llc Methods and systems for detecting cracks in electronic devices
US10269110B2 (en) 2016-06-28 2019-04-23 Ecoatm, Llc Methods and systems for detecting cracks in illuminated electronic device screens
US10401411B2 (en) 2014-09-29 2019-09-03 Ecoatm, Llc Maintaining sets of cable components used for wired analysis, charging, or other interaction with portable electronic devices
US10417615B2 (en) 2014-10-31 2019-09-17 Ecoatm, Llc Systems and methods for recycling consumer electronic devices
US10445708B2 (en) 2014-10-03 2019-10-15 Ecoatm, Llc System for electrically testing mobile devices at a consumer-operated kiosk, and associated devices and methods
US10475002B2 (en) 2014-10-02 2019-11-12 Ecoatm, Llc Wireless-enabled kiosk for recycling consumer devices
US10572946B2 (en) 2014-10-31 2020-02-25 Ecoatm, Llc Methods and systems for facilitating processes associated with insurance services and/or other services for electronic devices
WO2020049551A1 (en) 2018-09-06 2020-03-12 Orbotech Ltd. Multimodality multiplexed illumination for optical inspection systems
US10825082B2 (en) 2008-10-02 2020-11-03 Ecoatm, Llc Apparatus and method for recycling mobile phones
US10860990B2 (en) 2014-11-06 2020-12-08 Ecoatm, Llc Methods and systems for evaluating and recycling electronic devices
US11010841B2 (en) 2008-10-02 2021-05-18 Ecoatm, Llc Kiosk for recycling electronic devices
US11080672B2 (en) 2014-12-12 2021-08-03 Ecoatm, Llc Systems and methods for recycling consumer electronic devices
US20210350495A1 (en) * 2020-02-10 2021-11-11 Cognex Corporation Composite three-dimensional blob tool and method for operating the same
US20220292665A1 (en) * 2019-10-02 2022-09-15 Konica Minolta, Inc. Workpiece surface defect detection device and detection method, workpiece surface inspection system, and program
US11462868B2 (en) 2019-02-12 2022-10-04 Ecoatm, Llc Connector carrier for electronic device kiosk
US11482067B2 (en) 2019-02-12 2022-10-25 Ecoatm, Llc Kiosk for evaluating and purchasing used electronic devices
US20230084949A1 (en) * 2021-09-13 2023-03-16 Quanta Computer Inc. Systems and methods for monitoring antenna arrays
US11798250B2 (en) 2019-02-18 2023-10-24 Ecoatm, Llc Neural network based physical condition evaluation of electronic devices, and associated systems and methods
US11922467B2 (en) 2020-08-17 2024-03-05 ecoATM, Inc. Evaluating an electronic device using optical character recognition
US11974046B2 (en) 2019-08-27 2024-04-30 Orbotech Ltd. Multimodality multiplexed illumination for optical inspection systems

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5822055A (en) * 1995-06-06 1998-10-13 Kla Instruments Corporation Optical inspection of a specimen using multi-channel responses from the specimen using bright and darkfield detection
US5946029A (en) * 1996-06-25 1999-08-31 Matsushita Electric Works, Ltd Image processing process
US5982493A (en) * 1998-06-02 1999-11-09 Motorola, Inc. Apparatus and method for acquiring multiple images
US5995220A (en) * 1998-02-19 1999-11-30 Komatsu, Ltd. Semiconductor package inspection apparatus
US6075883A (en) * 1996-11-12 2000-06-13 Robotic Vision Systems, Inc. Method and system for imaging an object or pattern

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS61293657A (en) * 1985-06-21 1986-12-24 Matsushita Electric Works Ltd Method for inspecting soldering appearance
JP2953736B2 (en) * 1990-03-20 1999-09-27 松下電器産業株式会社 Solder shape inspection method
US5064291A (en) * 1990-04-03 1991-11-12 Hughes Aircraft Company Method and apparatus for inspection of solder joints utilizing shape determination from shading
JP3072998B2 (en) * 1990-04-18 2000-08-07 株式会社日立製作所 Soldering condition inspection method and apparatus
DE19725633C1 (en) * 1997-06-17 1998-12-17 Zentrum Fuer Neuroinformatik G Method and arrangement for analyzing the nature of a surface

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5822055A (en) * 1995-06-06 1998-10-13 Kla Instruments Corporation Optical inspection of a specimen using multi-channel responses from the specimen using bright and darkfield detection
US5946029A (en) * 1996-06-25 1999-08-31 Matsushita Electric Works, Ltd Image processing process
US6075883A (en) * 1996-11-12 2000-06-13 Robotic Vision Systems, Inc. Method and system for imaging an object or pattern
US5995220A (en) * 1998-02-19 1999-11-30 Komatsu, Ltd. Semiconductor package inspection apparatus
US5982493A (en) * 1998-06-02 1999-11-09 Motorola, Inc. Apparatus and method for acquiring multiple images

Cited By (129)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060006311A1 (en) * 2001-06-22 2006-01-12 Orbotech Ltd High-sensitivity optical scanning using memory integration
US7129509B2 (en) * 2001-06-22 2006-10-31 Orbotech, Ltd. High-sensitivity optical scanning using memory integration
US8536506B2 (en) 2001-06-22 2013-09-17 Orbotech Ltd. Imaging device and method for high-sensitivity optical scanning and integrated circuit therefor
US8119969B2 (en) 2001-06-22 2012-02-21 Orbotech Ltd Imaging device and method for high-sensitivity optical scanning and integrated circuit therefor
US9232114B2 (en) 2001-06-22 2016-01-05 Orbotech Ltd. Imaging device and method for high-sensitivity optical scanning and integrated circuit therefor
US7272255B2 (en) * 2002-01-15 2007-09-18 Nec Corporation Three-dimensional body photography apparatus, three-dimensional geometricmodel generation apparatus, three-dimensional geometric model generation method, and record medium in which three-dimensional geometric model generation program was stored
US20030169917A1 (en) * 2002-01-15 2003-09-11 Nec Corporation Three-dimensional body photography apparatus, three-dimensional geometricmodel generation apparatus, three-dimensional geometric model generation method, and record medium in which three-dimensional geometric model generation program was stored
US20040168881A1 (en) * 2003-01-10 2004-09-02 National Rejectors Inc. Gmbh Process for identifying an embossed image of a coin in an automatice coin machine
US7004299B2 (en) * 2003-01-10 2006-02-28 National Rejectors, Inc. Gmbh Process for identifying an embossed image of a coin in an automatic coin machine
US20040245334A1 (en) * 2003-06-06 2004-12-09 Sikorski Steven Maurice Inverted terminal presentation scanner and holder
US20050103753A1 (en) * 2003-11-18 2005-05-19 Matsushita Electric Industrial Co., Ltd. Light processing apparatus
US7532749B2 (en) * 2003-11-18 2009-05-12 Panasonic Corporation Light processing apparatus
US7646906B2 (en) 2004-01-29 2010-01-12 Kla-Tencor Technologies Corp. Computer-implemented methods for detecting defects in reticle design data
US20110221673A1 (en) * 2004-04-06 2011-09-15 Symbol Technologies, Inc. System and method for monitoring a mobile computing product/arrangement
US8773260B2 (en) 2004-04-06 2014-07-08 Symbol Technologies, Inc. System and method for monitoring a mobile computing product/arrangement
US20110205376A1 (en) * 2004-04-06 2011-08-25 Symbol Technologies, Inc. System and method for monitoring a mobile computing product/arrangement
US20100100623A1 (en) * 2004-04-06 2010-04-22 Thomas Wulff System and method for monitoring a mobile computing product/arrangement
US20110205076A1 (en) * 2004-04-06 2011-08-25 Symbol Technologies, Inc. System and method for monitoring a mobile compputing product/arrangement
US7551765B2 (en) * 2004-06-14 2009-06-23 Delphi Technologies, Inc. Electronic component detection system
US20050286754A1 (en) * 2004-06-14 2005-12-29 Thomas Steven P Electronic component detection system
US7689966B2 (en) 2004-09-14 2010-03-30 Kla-Tencor Technologies Corp. Methods, systems, and carrier media for evaluating reticle layout data
US8111900B2 (en) 2004-12-07 2012-02-07 Kla-Tencor Technologies Corp. Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle
US20060291714A1 (en) * 2004-12-07 2006-12-28 Kenong Wu Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle
US20100226562A1 (en) * 2004-12-07 2010-09-09 Kla-Tencor Technologies Corporation Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle
US7729529B2 (en) * 2004-12-07 2010-06-01 Kla-Tencor Technologies Corp. Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle
US20070076943A1 (en) * 2005-04-15 2007-04-05 Vistec Semiconductor Systems Jena Gmbh Method and apparatus for inspecting a wafer
GB2427541A (en) * 2005-06-24 2007-01-03 Aew Delford Group Ltd Vision system for food cutting and portioning apparatus
GB2428182A (en) * 2005-06-24 2007-01-24 Aew Delford Systems Vision system for food cutting and portioning apparatus
US7822513B2 (en) 2005-07-27 2010-10-26 Symbol Technologies, Inc. System and method for monitoring a mobile computing product/arrangement
US20070027585A1 (en) * 2005-07-27 2007-02-01 Thomas Wulff System and method for monitoring a mobile computing product/arrangement
US7769225B2 (en) * 2005-08-02 2010-08-03 Kla-Tencor Technologies Corp. Methods and systems for detecting defects in a reticle design pattern
US7676077B2 (en) 2005-11-18 2010-03-09 Kla-Tencor Technologies Corp. Methods and systems for utilizing design data in combination with inspection data
US8139843B2 (en) 2005-11-18 2012-03-20 Kla-Tencor Technologies Corp. Methods and systems for utilizing design data in combination with inspection data
US8923600B2 (en) 2005-11-18 2014-12-30 Kla-Tencor Technologies Corp. Methods and systems for utilizing design data in combination with inspection data
US8041103B2 (en) 2005-11-18 2011-10-18 Kla-Tencor Technologies Corp. Methods and systems for determining a position of inspection data in design data space
US20070298751A1 (en) * 2006-06-21 2007-12-27 Thomas Wulff System and method for monitoring a mobile device
US8594742B2 (en) 2006-06-21 2013-11-26 Symbol Technologies, Inc. System and method for monitoring a mobile device
US20070297028A1 (en) * 2006-06-21 2007-12-27 Thomas Wulff System and device for monitoring a computing device
US7877722B2 (en) 2006-12-19 2011-01-25 Kla-Tencor Corp. Systems and methods for creating inspection recipes
US8194968B2 (en) 2007-01-05 2012-06-05 Kla-Tencor Corp. Methods and systems for using electrical information for a device being fabricated on a wafer to perform one or more defect-related functions
US7962863B2 (en) 2007-05-07 2011-06-14 Kla-Tencor Corp. Computer-implemented methods, systems, and computer-readable media for determining a model for predicting printability of reticle features on a wafer
US7738093B2 (en) 2007-05-07 2010-06-15 Kla-Tencor Corp. Methods for detecting and classifying defects on a reticle
US8213704B2 (en) 2007-05-09 2012-07-03 Kla-Tencor Corp. Methods and systems for detecting defects in a reticle design pattern
US8204296B2 (en) 2007-07-20 2012-06-19 Kla-Tencor Corp. Methods for generating a standard reference die for use in a die to standard reference die inspection and methods for inspecting a wafer
US7796804B2 (en) 2007-07-20 2010-09-14 Kla-Tencor Corp. Methods for generating a standard reference die for use in a die to standard reference die inspection and methods for inspecting a wafer
US7711514B2 (en) 2007-08-10 2010-05-04 Kla-Tencor Technologies Corp. Computer-implemented methods, carrier media, and systems for generating a metrology sampling plan
US7975245B2 (en) 2007-08-20 2011-07-05 Kla-Tencor Corp. Computer-implemented methods for determining if actual defects are potentially systematic defects or potentially random defects
US8139844B2 (en) 2008-04-14 2012-03-20 Kla-Tencor Corp. Methods and systems for determining a defect criticality index for defects on wafers
US9659670B2 (en) 2008-07-28 2017-05-23 Kla-Tencor Corp. Computer-implemented methods, computer-readable media, and systems for classifying defects detected in a memory device area on a wafer
US10853873B2 (en) 2008-10-02 2020-12-01 Ecoatm, Llc Kiosks for evaluating and purchasing used electronic devices and related technology
US11935138B2 (en) 2008-10-02 2024-03-19 ecoATM, Inc. Kiosk for recycling electronic devices
US9881284B2 (en) 2008-10-02 2018-01-30 ecoATM, Inc. Mini-kiosk for recycling electronic devices
US10055798B2 (en) 2008-10-02 2018-08-21 Ecoatm, Llc Kiosk for recycling electronic devices
US11107046B2 (en) 2008-10-02 2021-08-31 Ecoatm, Llc Secondary market and vending system for devices
US9818160B2 (en) 2008-10-02 2017-11-14 ecoATM, Inc. Kiosk for recycling electronic devices
US11790328B2 (en) 2008-10-02 2023-10-17 Ecoatm, Llc Secondary market and vending system for devices
US11907915B2 (en) 2008-10-02 2024-02-20 Ecoatm, Llc Secondary market and vending system for devices
US11443289B2 (en) 2008-10-02 2022-09-13 Ecoatm, Llc Secondary market and vending system for devices
US10032140B2 (en) * 2008-10-02 2018-07-24 ecoATM, LLC. Systems for recycling consumer electronic devices
US11526932B2 (en) 2008-10-02 2022-12-13 Ecoatm, Llc Kiosks for evaluating and purchasing used electronic devices and related technology
US10825082B2 (en) 2008-10-02 2020-11-03 Ecoatm, Llc Apparatus and method for recycling mobile phones
US11080662B2 (en) 2008-10-02 2021-08-03 Ecoatm, Llc Secondary market and vending system for devices
US10157427B2 (en) 2008-10-02 2018-12-18 Ecoatm, Llc Kiosk for recycling electronic devices
US20130144797A1 (en) * 2008-10-02 2013-06-06 ecoATM, Inc. Method And Apparatus For Recycling Electronic Devices
US11010841B2 (en) 2008-10-02 2021-05-18 Ecoatm, Llc Kiosk for recycling electronic devices
US9904911B2 (en) 2008-10-02 2018-02-27 ecoATM, Inc. Secondary market and vending system for devices
US8775101B2 (en) 2009-02-13 2014-07-08 Kla-Tencor Corp. Detecting defects on a wafer
US8204297B1 (en) 2009-02-27 2012-06-19 Kla-Tencor Corp. Methods and systems for classifying defects detected on a reticle
US8112241B2 (en) 2009-03-13 2012-02-07 Kla-Tencor Corp. Methods and systems for generating an inspection process for a wafer
US20100245813A1 (en) * 2009-03-24 2010-09-30 Orbotech Ltd. Multi-modal imaging
US8144973B2 (en) 2009-03-24 2012-03-27 Orbotech Ltd. Multi-modal imaging
JP2011175600A (en) * 2010-02-25 2011-09-08 Canon Inc Recognition apparatus and method thereof, and computer program
US9087265B2 (en) * 2010-02-25 2015-07-21 Canon Kabushiki Kaisha Recognition apparatus and method thereof, and computer program
US20110206237A1 (en) * 2010-02-25 2011-08-25 Canon Kabushiki Kaisha Recognition apparatus and method thereof, and computer program
US8781781B2 (en) 2010-07-30 2014-07-15 Kla-Tencor Corp. Dynamic care areas
US20120031975A1 (en) * 2010-08-04 2012-02-09 The Code Corporation Illumination blocks for a graphical code reader
US9170211B2 (en) 2011-03-25 2015-10-27 Kla-Tencor Corp. Design-based inspection using repeating structures
US9087367B2 (en) 2011-09-13 2015-07-21 Kla-Tencor Corp. Determining design coordinates for wafer defects
US8831334B2 (en) 2012-01-20 2014-09-09 Kla-Tencor Corp. Segmentation for wafer inspection
US8826200B2 (en) 2012-05-25 2014-09-02 Kla-Tencor Corp. Alteration for wafer inspection
CN103674974A (en) * 2012-09-06 2014-03-26 株式会社岛津制作所 Inspection apparatus
US9189844B2 (en) 2012-10-15 2015-11-17 Kla-Tencor Corp. Detecting defects on a wafer using defect-specific information
US9053527B2 (en) 2013-01-02 2015-06-09 Kla-Tencor Corp. Detecting defects on a wafer
US9134254B2 (en) 2013-01-07 2015-09-15 Kla-Tencor Corp. Determining a position of inspection system output in design data space
US9311698B2 (en) 2013-01-09 2016-04-12 Kla-Tencor Corp. Detecting defects on a wafer using template image matching
US9092846B2 (en) 2013-02-01 2015-07-28 Kla-Tencor Corp. Detecting defects on a wafer using defect-specific and multi-channel information
US9865512B2 (en) 2013-04-08 2018-01-09 Kla-Tencor Corp. Dynamic design attributes for wafer inspection
US9310320B2 (en) 2013-04-15 2016-04-12 Kla-Tencor Corp. Based sampling and binning for yield critical defects
JP2016035405A (en) * 2014-08-01 2016-03-17 リコーエレメックス株式会社 Image inspection device, image inspection system, and image inspection method
US10401411B2 (en) 2014-09-29 2019-09-03 Ecoatm, Llc Maintaining sets of cable components used for wired analysis, charging, or other interaction with portable electronic devices
US10438174B2 (en) 2014-10-02 2019-10-08 Ecoatm, Llc Application for device evaluation and other processes associated with device recycling
US11126973B2 (en) 2014-10-02 2021-09-21 Ecoatm, Llc Wireless-enabled kiosk for recycling consumer devices
US10496963B2 (en) 2014-10-02 2019-12-03 Ecoatm, Llc Wireless-enabled kiosk for recycling consumer devices
US11790327B2 (en) 2014-10-02 2023-10-17 Ecoatm, Llc Application for device evaluation and other processes associated with device recycling
US9911102B2 (en) 2014-10-02 2018-03-06 ecoATM, Inc. Application for device evaluation and other processes associated with device recycling
US10475002B2 (en) 2014-10-02 2019-11-12 Ecoatm, Llc Wireless-enabled kiosk for recycling consumer devices
US11734654B2 (en) 2014-10-02 2023-08-22 Ecoatm, Llc Wireless-enabled kiosk for recycling consumer devices
US10445708B2 (en) 2014-10-03 2019-10-15 Ecoatm, Llc System for electrically testing mobile devices at a consumer-operated kiosk, and associated devices and methods
US11232412B2 (en) 2014-10-03 2022-01-25 Ecoatm, Llc System for electrically testing mobile devices at a consumer-operated kiosk, and associated devices and methods
US10417615B2 (en) 2014-10-31 2019-09-17 Ecoatm, Llc Systems and methods for recycling consumer electronic devices
US10572946B2 (en) 2014-10-31 2020-02-25 Ecoatm, Llc Methods and systems for facilitating processes associated with insurance services and/or other services for electronic devices
US11436570B2 (en) 2014-10-31 2022-09-06 Ecoatm, Llc Systems and methods for recycling consumer electronic devices
US10860990B2 (en) 2014-11-06 2020-12-08 Ecoatm, Llc Methods and systems for evaluating and recycling electronic devices
US11080672B2 (en) 2014-12-12 2021-08-03 Ecoatm, Llc Systems and methods for recycling consumer electronic devices
US11315093B2 (en) 2014-12-12 2022-04-26 Ecoatm, Llc Systems and methods for recycling consumer electronic devices
US20160267339A1 (en) * 2015-03-11 2016-09-15 Canon Kabushiki Kaisha Image processing apparatus and method of generating face image
US10558851B2 (en) * 2015-03-11 2020-02-11 Canon Kabushiki Kaisha Image processing apparatus and method of generating face image
JP2017032548A (en) * 2015-06-26 2017-02-09 コグネックス・コーポレイション Using 3d vision for automated industrial inspection
JP7037876B2 (en) 2015-06-26 2022-03-17 コグネックス・コーポレイション Use of 3D vision in automated industrial inspection
TWI626438B (en) * 2015-09-30 2018-06-11 佳能股份有限公司 Inspection apparatus and article manufacturing method
JP2017122682A (en) * 2016-01-08 2017-07-13 株式会社キーエンス Visual inspection device, visual inspection method, and computer program executable by controller used in visual inspection device
US10127647B2 (en) 2016-04-15 2018-11-13 Ecoatm, Llc Methods and systems for detecting cracks in electronic devices
US9885672B2 (en) 2016-06-08 2018-02-06 ecoATM, Inc. Methods and systems for detecting screen covers on electronic devices
US10909673B2 (en) 2016-06-28 2021-02-02 Ecoatm, Llc Methods and systems for detecting cracks in illuminated electronic device screens
US11803954B2 (en) 2016-06-28 2023-10-31 Ecoatm, Llc Methods and systems for detecting cracks in illuminated electronic device screens
US10269110B2 (en) 2016-06-28 2019-04-23 Ecoatm, Llc Methods and systems for detecting cracks in illuminated electronic device screens
EP3847446A4 (en) * 2018-09-06 2022-06-01 Orbotech Ltd. Multimodality multiplexed illumination for optical inspection systems
WO2020049551A1 (en) 2018-09-06 2020-03-12 Orbotech Ltd. Multimodality multiplexed illumination for optical inspection systems
US11482067B2 (en) 2019-02-12 2022-10-25 Ecoatm, Llc Kiosk for evaluating and purchasing used electronic devices
US11462868B2 (en) 2019-02-12 2022-10-04 Ecoatm, Llc Connector carrier for electronic device kiosk
US11843206B2 (en) 2019-02-12 2023-12-12 Ecoatm, Llc Connector carrier for electronic device kiosk
US11798250B2 (en) 2019-02-18 2023-10-24 Ecoatm, Llc Neural network based physical condition evaluation of electronic devices, and associated systems and methods
US11974046B2 (en) 2019-08-27 2024-04-30 Orbotech Ltd. Multimodality multiplexed illumination for optical inspection systems
US20220292665A1 (en) * 2019-10-02 2022-09-15 Konica Minolta, Inc. Workpiece surface defect detection device and detection method, workpiece surface inspection system, and program
US20210350495A1 (en) * 2020-02-10 2021-11-11 Cognex Corporation Composite three-dimensional blob tool and method for operating the same
US11966996B2 (en) * 2020-02-10 2024-04-23 Cognex Corporation Composite three-dimensional blob tool and method for operating the same
US11922467B2 (en) 2020-08-17 2024-03-05 ecoATM, Inc. Evaluating an electronic device using optical character recognition
US20230084949A1 (en) * 2021-09-13 2023-03-16 Quanta Computer Inc. Systems and methods for monitoring antenna arrays
US11742962B2 (en) * 2021-09-13 2023-08-29 Quanta Computer Inc. Systems and methods for monitoring antenna arrays

Also Published As

Publication number Publication date
WO2002099403A1 (en) 2002-12-12

Similar Documents

Publication Publication Date Title
US20020186878A1 (en) System and method for multiple image analysis
US20220084183A1 (en) Defect detection device, defect detection method, and program
US9552636B2 (en) Detecting defects on a wafer using defect-specific and multi-channel information
US6888959B2 (en) Method of inspecting a semiconductor device and an apparatus thereof
US7177458B1 (en) Reduction of false alarms in PCB inspection
CA2638415C (en) Patterned wafer defect inspection system and method
US7420542B2 (en) Apparatus for capturing and analyzing light and method embodied therein
US20030117616A1 (en) Wafer external inspection apparatus
US7024031B1 (en) System and method for inspection using off-angle lighting
JPH06307833A (en) Surface unevenness shape recognizing device
JP2002296192A (en) Method for inspecting flaw using color illumination
JP2009264882A (en) Visual inspection device
JP2002250700A (en) Method and device for inspecting pattern
KR102015620B1 (en) System and Method for detecting Metallic Particles
JPH11352073A (en) Foreign matter inspection method and apparatus thereof
Katafuchi et al. A method for inspecting industrial parts surfaces based on an optics model
JP2004212218A (en) Sample inspection method and inspection apparatus
JPS61193007A (en) Inspecting method for rod type projection body
JP2003057193A (en) Foreign matter checking apparatus
US20240094145A1 (en) Detection method and system for determining the location of a surface defect on a front or back surface of a transparent film
JPH0413953A (en) Detect inspection preprocessor for electronic component molded form
JP2532513B2 (en) Object presence inspection method
KR100564871B1 (en) Inspecting method and apparatus for repeated micro-miniature patterns
Kim 3-Dimensional Micro Solder Ball Inspection Using LED Reflection Image
CN113447485A (en) Optical detection method

Legal Events

Date Code Title Description
AS Assignment

Owner name: ASTI HOLDINGS, INC., SINGAPORE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HOON, TAN SEOW;RAO, SREENIVAS;REEL/FRAME:012439/0766;SIGNING DATES FROM 20010829 TO 20010921

AS Assignment

Owner name: ASTI HOLDINGS LIMITED, SINGAPORE

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE RECEIVING PARTY ADDRESS PREVIOUSLY RECORDED ON REEL 012429 FRAME 0766;ASSIGNORS:HOON, TAN SEOW;SREENIVAS, RAO;REEL/FRAME:012968/0341;SIGNING DATES FROM 20010829 TO 20010921

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION