US20020186878A1 - System and method for multiple image analysis - Google Patents
System and method for multiple image analysis Download PDFInfo
- Publication number
- US20020186878A1 US20020186878A1 US09/876,795 US87679501A US2002186878A1 US 20020186878 A1 US20020186878 A1 US 20020186878A1 US 87679501 A US87679501 A US 87679501A US 2002186878 A1 US2002186878 A1 US 2002186878A1
- Authority
- US
- United States
- Prior art keywords
- image data
- component
- data
- receiving
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/95—Investigating the presence of flaws or contamination characterised by the material or shape of the object to be examined
- G01N21/956—Inspecting patterns on the surface of objects
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
- G01B11/245—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures using a plurality of fixed, simultaneously operating transducers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
- G06T7/001—Industrial image inspection using an image reference approach
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
- G06T7/586—Depth or shape recovery from multiple images from multiple light sources, e.g. photometric stereo
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
- G06T7/62—Analysis of geometric attributes of area, perimeter, diameter or volume
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
- G06T2207/30148—Semiconductor; IC; Wafer
Definitions
- the present invention pertains to the field of semiconductor devices, and more particularly to a system and method for inspecting semiconductor devices that uses multiple two-dimensional images to generate third dimension data.
- Image data analysis systems for inspecting semiconductor components are known in the art. Such image data analysis systems attempt to determine the state of the semiconducting component or other inspected components by analyzing image data, which is typically comprised of an N ⁇ M array of picture elements or “pixels.” The brightness value of each pixel of a test image is typically compared to the brightness element of a corresponding pixel of a reference image, and the comparison data is analyzed to determine whether or not unacceptable defects exist on the semiconducting device, component or other object being inspected. For example, image data analysis is used to determine whether the variation in the dimensions of an element of the component exceed allowable tolerances for such dimensions.
- One drawback with known image data inspection systems is the difficulty in determining the three-dimensional nature of elements. Such image data is typically taken from a single angle, such that any three-dimensional aspects or flaws may be difficult to detect.
- a common method for determining the three-dimensional aspects of a semiconductor device or component that is being inspected is to use a laser beam to trace a line, and to determine when the line varies from a straight line, where such variations are then correlated to defects in the semiconducting device or component.
- the semiconducting device or component contains a large number of elements, it is necessary to trace a laser line through each of the elements, which can require movement of the component to a number of different locations. Likewise, it is possible that the laser drawn line may not lie on a defect, such that the defect could be missed.
- a system and method for multiple image analysis are provided that use image data generated by illuminating a component from two or more lighting angles, which allows three-dimensional aspects of the component to be determined.
- a system for analyzing multiple images is provided, such as to locate defects in a test component.
- the system includes a first light source, such as one that emits blue light, and a second light source, such as one that emits red light.
- the system also includes a camera, where the camera and the light sources are focused on an area where a test piece is to be placed.
- a multiple image processor is connected to the first light source, the second light source, and the camera.
- the multiple image processor causes the first light source and the second light source to turn on, such as in sequence, and also causes the camera to generate two or more sets of image data, such as one set when each of the light sources is illuminated, through the use of filters or tuned pixels, or otherwise.
- the present invention provides many important technical advantages.
- One important technical advantage of the present invention is a system and method for multiple image analysis that uses two or more sets of image data to analyze a component. Each set of image data is obtained when the component is illuminated by a light source having a different lighting angle, which creates shaded areas that can be analyzed to determine whether they indicate the existence of damage or unacceptable dimensional variations on the component.
- FIG. 1 is a diagram of a system for performing multiple image analysis in accordance with an exemplary embodiment of the present invention
- FIGS. 2A, 2B, and 2 C show an exemplary undamaged element and corresponding bright and shaded regions generated by illumination from light sources;
- FIGS. 3A, 3B, and 3 C show an exemplary damaged element and corresponding bright and shaded regions generated by illumination from light sources
- FIG. 4 is a diagram of a system for processing image data from multiple images in accordance with an exemplary embodiment of the present invention
- FIG. 5 is a flowchart of a method for analyzing image data from multiple images in accordance with an exemplary embodiment of the present invention
- FIG. 6 is a flowchart of a method for analyzing image data in accordance with an exemplary embodiment of the present invention.
- FIG. 7 is a flowchart of a method for performing image data analysis for multiple images in accordance with an exemplary embodiment of the present invention.
- FIG. 1 is a diagram of a system 100 for performing multiple image analysis in accordance with an exemplary embodiment of the present invention.
- System 100 allows three-dimensional aspects of an inspected device or component to be determined from images obtained from two or more different viewing angles.
- System 100 includes multiple image processor 102 , which can be implemented in hardware, software, or a suitable combination of hardware and software, and which can be one or more software systems operating on a general purpose processing platform.
- a software system can include one or more objects, agents, subroutines, lines of code, threads, two or more lines of code or other suitable software structures operating in two or more separate software applications, or other suitable software structures, and can operate on two or more different processors, or other suitable configurations of processors.
- a software system can include one or more lines of code or other software structures operating in a general purpose software application, such as an operating system, and one or more lines of code or other suitable software structures operating in a specific purpose software application.
- Coupled to light sources 104 a and 104 b can include a physical connection (such as a copper conductor), a virtual connection (such as through one or more randomly assigned data memory locations of a data memory device), a logical connection (such as through one or more logical gates of a semiconducting device), a wireless connection, other suitable connections, or a suitable combination of such connections.
- systems and components are coupled to other systems and components through intervening systems and components, such as through an operating system of a general purpose processor platform.
- Multiple image processor 102 is also coupled to camera 106 .
- Camera 106 can be a charge coupled device (CCD), a CMOS imaging device, or other suitable imaging devices that are focused on a component 108 having a plurality of elements 110 .
- Light sources 104 a and 104 b are also focused on component 108 , and illuminate component 108 from different angles as shown in FIG. 1.
- the light illuminating component 108 from light source 104 a will create shaded regions that are different from the shaded regions created by light illuminating component 108 from light source 104 b .
- Additional light sources can be used where suitable to create additional shaded regions.
- Camera 106 is used to record image data of component 108 while it is being illuminated by light sources 104 a and 104 b .
- camera 106 is controlled by multiple image processor 102 to store a first set of image data of component 108 when light source 104 a is on, and to store a second set of image data when light source 104 b is on.
- camera 106 can store image data when both of light sources 104 a and 104 b are on, such as when the light sources use different frequencies of light.
- camera 106 can record image data according to the frequency of the light that creates the image, such as by including one or more light filters, two or more sets of pixels that are tuned to received predetermined frequencies of light, or to otherwise differentiate between light illuminated from light sources 104 a and 104 b , such that multiple sets of image data can be concurrently gathered.
- a component 108 is placed in the focal area of light sources 104 a and 104 b and camera 106 for inspection.
- Multiple image processor 102 then causes component 108 to be illuminated and causes camera 106 to produce image data, such as by generating an N ⁇ M array of pixels of image data, which can then be stored by multiple image processor 102 or other suitable storage systems or devices. Because of the angular difference between light sources 104 a and 104 b relative to component 108 , shaded regions are generated from elements 110 .
- Multiple image processor 102 can analyze these shaded regions to determine whether they are indicative of any damage or defects to component 108 , elements 110 , or other suitable indications.
- multiple image processor 102 can determine whether three-dimensional defects or other variations in component 108 or elements 110 exist. For example, if one of elements 110 is damaged, then the shaded regions generated by that element 110 when it is illuminated by light sources 104 a and 104 b will vary from the shaded regions generated for undamaged reference images. Furthermore, the variations in pixel brightness between corresponding pixels of the test image data and the reference image data, as illuminated by light sources 104 a and 104 b , can also be analyzed to generate an approximation of differences in height, dimensions, or other data that can be used to approximate a three-dimensional analysis.
- FIGS. 2A, 2B, and 2 C show an exemplary undamaged element 110 and corresponding bright and shaded regions generated by illumination from light sources 104 a and 104 b (not explicitly shown).
- FIG. 2A shows an exemplary undamaged element 110 , which is semi-spherical in shape.
- the circular outline of element 110 as viewed from overhead is shown with an illuminated region and a shaded region corresponding to the shadow generated by light source 104 a .
- the shaded region generates a distinctive pattern which is indicative of a spherical configuration of element 110 .
- the shaded region of element 110 is on the opposite face, as a result of the location of light source 104 b .
- the shaded regions generated shown in FIGS. 2 b and 2 c can be used as a reference for an undamaged element 110 .
- the differences in pixel brightness data between FIG. 2B and FIG. 2C and the known angle of illumination from light sources 104 a and 104 b can also be used to estimate the dimensional variations of element 110 .
- it can be determined from areas in FIG. 2B and FIG. 2C in which the pixel brightness data is a maximum and does not vary that such areas are not directly blocked from direct exposure by either source.
- an obstruction is blocking those pixels, and that the obstruction is located between light source having the lower brightness values and the location of the pixel being analyzed.
- Other suitable procedures can be used to estimate the size and location of dimensional variations based upon pixel data, such as the use of empirically developed pass/fail ratios based upon the size of areas in which pixel brightness variations between two or more images exceed predetermined levels.
- FIGS. 3A, 3B, and 3 C show an exemplary damaged element 110 and corresponding bright and shaded regions generated by illumination from light sources 104 a and 104 b (not explicitly shown).
- FIG. 3A shows the damaged element 110 which varies from semi-circular, such as by an indentation. As shown in FIG. 3B, this indentation creates shaded regions 302 and 304 . These shaded regions 302 and 304 are different from shaded region 202 in FIG. 2 b . These exemplary variations can be used to detect three-dimensional variations in element 110 that would otherwise be difficult to detect from a single image, depending on the angle of illumination.
- FIG. 3C includes shaded region 306 which varies from shaded region 204 . The pixels defining these regions can be compared between a test image, such as that shown in FIG. 3B and FIG. 3C, and a reference image, such as that shown in FIG. 2B and FIG.
- the composite images formed by combining image data from shaded regions 202 and 204 with FIGS. 3B and 3C can be used and compared, so as to generate additional comparison points.
- the variations in pixel brightness between the reference images and test images can also be used, in conjunction with the known angular position of light sources, to estimate the location and size of obstructions, deformations, or other features.
- the shaded regions generated by illumination from light sources 104 a and 104 b can be used to generate three-dimensional image data from two dimensional image data.
- Pixel brightness data can also be used to estimate the dimensional variations between a test image and a reference image.
- FIG. 4 is a diagram of a system 400 for processing image data from multiple images in accordance with an exemplary embodiment of the present invention.
- System 400 includes multiple image processor 102 and light sequence controller 402 , first image analyzer 404 , second image analyzer 406 , image comparator 408 , and 3D image constructor 410 , each of which can be implemented in hardware, software, or a suitable combination of hardware and software, and which can be one or more software systems operating on a general purpose processor platform.
- Light sequence controller 402 controls the sequence in which light sources 104 a , 104 b , and other suitable lights illuminate a component 108 . Likewise, light sequence controller 402 also controls the operation of camera 106 , such that when a first light source is illuminating the component 108 , camera 106 captures first image data, and when a second light source is illuminating the component 108 , camera 106 captures or generates second image data. Likewise, light sequence controller 402 can control light sources having different frequencies, such that camera 106 can generate multiple sets of image data simultaneously so as to decrease the amount of time required to generate the multiple sets of image data.
- First image analyzer 404 and second image analyzer 406 receive an N ⁇ M array of pixels of brightness data, and analyze the pixel data to determine whether the pixel data is acceptable, requires additional analysis such as comparison with a reference image or dimensional analysis, or is unacceptable. First image analyzer 404 and second image analyzer 406 then generate status data indicating whether the pixel data is acceptable, requires further analysis, or is unacceptable. In one exemplary embodiment, first image analyzer 404 receives pixel array data generated when light source 104 a illuminates component 108 and second image analyzer 406 receives pixel array data generated when light source 104 b illuminates component 108 . Additional image analyzers can also be used to accommodate light sources illuminating the component 108 from different angles. First image analyzer 404 and second image analyzer 406 perform pixel brightness analysis of the corresponding images.
- first image analyzer 404 and second image analyzer 406 determine whether the pixel data indicates that the number and magnitude of variations in pixel brightness data exceed predetermined maximum allowable numbers and magnitudes, such that it is determinable whether the component contains unacceptable dimensional variations without additional image data analysis. In addition, first image analyzer 404 and second image analyzer 406 can determine whether the pixel data falls within a range of values that indicates that further analysis is required.
- Image comparator 408 receives first image data and second image data and generates difference image data, such as by subtracting pixel brightness data for corresponding pixels between a first image and a second image.
- Image comparator 408 can perform comparator analysis of first test image data and first reference image data, second test image data and second reference image data, composite test image data and composite reference image data, or other suitable sets of corresponding image data.
- Image comparator 408 can also generate absolute brightness variation data, relative brightness variation data, or other suitable brightness variation data.
- 3D image constructor 410 can receive the test image data, reference image data, difference image data, composite image data, or other suitable image data and determine whether defects, variations, or other features of element 110 or other elements exceed allowable variations for such elements.
- 3D image constructor 410 can determine from the known angle of illumination of light sources 104 a , 104 b and other light sources, and from the brightness values of pixels generated when such light sources illuminate the component, whether the light source is illuminating the feature or element 110 at that corresponding position.
- 3D image constructor 410 can include predetermined ranges for allowable variations, such as histogram data, pixel area mapping data, and other suitable data.
- 3D image constructor 410 can be used to generate dimensional variation data after determining whether a variation or feature in an element 110 exceeds allowable limits, such that the component having the element can be rejected in the event the damage or dimensional variation in the element 110 exceeds such limits.
- system 400 is used to control the inspection of a component, to generate test image data, to analyze the test image data, and to estimate three-dimensional variations or features of a test image.
- System 400 utilizes image data generated by illuminating the component from two or more angles, can combine the test image data and compare the test image data to reference image data, and can process any difference image data to make determinations on whether or not to accept or reject a component.
- FIG. 5 is a flowchart of a method 500 for analyzing image data from multiple images in accordance with an exemplary embodiment of the present invention.
- Method 500 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
- Method 500 begins at 502 where image data is obtained.
- the image data is obtained by simultaneously illuminating a component with multiple light sources from different angles, where each light source is illuminated at a different time.
- the light sources can provide light having different frequencies where the image data is generated at the same time and filters, tuned pixels, or other procedures are used to separate the image data created from each light source. The method then proceeds to 504 .
- each set of image data is analyzed.
- the sets of image data can be analyzed by generating histogram data showing the brightness of each pixel, by comparing each set of test image data to a set of reference image data and performing histogram analysis or other suitable analysis of the difference image data set, by combining the test image data and comparing the combined test image data to predetermined acceptable ranges for histogram data, by comparing the combined test image data to combined reference image data, or performing other suitable analyses.
- the method then proceeds to 506 .
- compared image data is obtained.
- an initial analysis is performed at 504 to determine whether additional analyses need to be performed, such that the analysis performed at 504 does not include comparator data.
- Other suitable processes can be used.
- the comparator image data is obtained for each single test image and reference image, composite test and reference images, or other suitable comparative data, the method proceeds to 512 .
- the image data and comparator data is analyzed to generate three-dimensional image data.
- the three-dimensional image data can include predetermined allowable ranges for three-dimensional variations that generate shaded regions of elements when illuminated by multiple light sources.
- the three-dimensional image data can include estimates of variations and components based upon the known angular relationship between the light sources and the component. The method then proceeds to 514 .
- the three-dimensional image data is applied to template data.
- the template data can include one or more templates that are used to estimate variations between measured brightness data and expected brightness data, so as to determine whether three-dimensional variations in the inspected component exceed allowable variations. The method then proceeds to 516 .
- the method proceeds to 518 where the image data is rejected, such as by rejecting the component, flagging the component for further manual inspection, or other suitable procedures. Otherwise, the method proceeds to 520 where the image data is accepted.
- method 500 is used to analyze multiple sets of image data for a test component in order to determine whether the component includes dimensional variations, damage, or other unacceptable condition.
- Method 500 further utilizes light sources having different angular relationships to the test component, where the known angular relationship of the light sources can be used in conjunction with the pixel brightness data to estimate 3-dimensional variations in the test component.
- FIG. 6 is a flowchart of a method 600 for analyzing image data in accordance with an exemplary embodiment of the present invention.
- Method 600 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
- Method 600 begins at 602 where a test piece is exposed to light from two different light frequencies and two different angular illumination zones. The method then proceeds to 604 and 608 in parallel.
- first image data is obtained, such as by filtering the light through a first filter, by using pixels tuned to the first light frequency, or other suitable methods.
- second image data is obtained, such as by filtering the light through a second filter, by using pixels tuned to a second light frequency, or other suitable methods.
- the method then proceeds to 606 from 604 and 610 from 608 , respectively.
- the pixel brightness variation data is analyzed for the first image data. For example, pixel histogram data can be generated and the variations in pixel brightness can be compared to predetermined acceptable ranges. Likewise, other suitable pixel brightness variation analysis methods can be used.
- similar pixel brightness variations are analyzed for the second image data. The method then proceeds to 612 and 614 , respectively.
- method 600 can be used to determine whether three-dimensional analysis of component image data should be performed, such as to perform a quick preliminary component image inspection analysis for the purpose of determining whether additional analysis should be performed.
- Method 600 also allows image data to be analyzed in parallel where suitable, such as when a parallel processor platform is being used to analyze the image data.
- method 600 can be implemented on a general purpose processing platform, such as through the use of multitasking or by otherwise simulating parallel processing.
- FIG. 7 is a flowchart of a method 700 for performing image data analysis for multiple images in accordance with an exemplary embodiment of the present invention.
- Method 700 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
- Method 700 begins at 702 and 704 in parallel.
- first reference image data is compared to first test image data
- second reference image data is compared to second test image data. This comparison can include a pixel to correlating pixel brightness subtraction to generate a difference image, or other suitable comparison procedures.
- the method then proceeds to 706 .
- a composite test image is formed.
- the composite test image can include two or more sets of image data generated from two or more different illumination angles, from two or more different light frequencies, or other suitable composite test data. The method then proceeds to 712 .
- the composite test image data is compared to composite reference image data, such as by performing a pixel to corresponding pixel subtraction or other suitable compare procedures. The method then proceeds to 714 .
- three-dimensional coordinates for the component being inspected are estimated from variations in the test image data as compared to the reference image data. For example, pixels at coordinates that have significant variations in brightness as a function of the angle of illumination can indicate the existence of an indentation, spur, bulge, or other deformity in the component. It may be determined by analysis, empirically, or otherwise that such variations in brightness that exceed certain levels correlate to dimensional variations. Likewise, an estimate of the dimensional variation can be calculated from the brightness data and the known angular position of each light source. Other suitable methods may also or alternatively be used. The method then proceeds to 716 .
- allowable variations can be determined empirically, by calculation, can be set according to a customer or industry standard, or through other suitable methods. The method then proceeds to 718 if it is determined that the variations exceed allowable ranges and the image data is rejected. Likewise, a message can be generated informing the operator that additional analysis or operator inspection is required. If it is determined at 716 that any variations in the image data are allowable then the method proceeds to 720 where the image data is accepted.
- method 700 allows a component to be inspected by illuminating the component from multiple light sources, such that the component generates shaded regions and bright regions.
- the shaded and bright regions of the component can be then analyzed and compared to reference image data to determine whether unacceptable variations or damage may exist on the component.
Abstract
A system for analyzing multiple images is provided, such as to locate defects in a test component. The system includes a first light source, such as one that emits blue light, and a second light source, such as one that emits red light. The system also includes a camera, where the camera and the light sources are focused on an area where a test piece is to be placed. A multiple image processor is connected to the first light source, the second light source, and the camera. The multiple image processor causes the first light source and the second light source to turn on, such as in sequence, and also causes the camera to generate two or more sets of image data, such as one set when each of the light sources is illuminated, through the use of filters or tuned pixels, or otherwise.
Description
- The present invention pertains to the field of semiconductor devices, and more particularly to a system and method for inspecting semiconductor devices that uses multiple two-dimensional images to generate third dimension data.
- Image data analysis systems for inspecting semiconductor components are known in the art. Such image data analysis systems attempt to determine the state of the semiconducting component or other inspected components by analyzing image data, which is typically comprised of an N×M array of picture elements or “pixels.” The brightness value of each pixel of a test image is typically compared to the brightness element of a corresponding pixel of a reference image, and the comparison data is analyzed to determine whether or not unacceptable defects exist on the semiconducting device, component or other object being inspected. For example, image data analysis is used to determine whether the variation in the dimensions of an element of the component exceed allowable tolerances for such dimensions.
- One drawback with known image data inspection systems is the difficulty in determining the three-dimensional nature of elements. Such image data is typically taken from a single angle, such that any three-dimensional aspects or flaws may be difficult to detect. For example, a common method for determining the three-dimensional aspects of a semiconductor device or component that is being inspected is to use a laser beam to trace a line, and to determine when the line varies from a straight line, where such variations are then correlated to defects in the semiconducting device or component. When the semiconducting device or component contains a large number of elements, it is necessary to trace a laser line through each of the elements, which can require movement of the component to a number of different locations. Likewise, it is possible that the laser drawn line may not lie on a defect, such that the defect could be missed.
- Thus, although it is known to perform analysis of image data of a component to determine whether variations in the dimensions of elements of the component exceed allowable tolerances, the determination of such dimensional variations in three dimensions is time-consuming and limited to small portions of the component.
- In accordance with the present invention, a system and method for multiple image analysis are provided that overcome known problems with analyzing image data.
- In particular, a system and method for multiple image analysis are provided that use image data generated by illuminating a component from two or more lighting angles, which allows three-dimensional aspects of the component to be determined.
- In accordance with an exemplary embodiment of the present invention, a system for analyzing multiple images is provided, such as to locate defects in a test component. The system includes a first light source, such as one that emits blue light, and a second light source, such as one that emits red light. The system also includes a camera, where the camera and the light sources are focused on an area where a test piece is to be placed. A multiple image processor is connected to the first light source, the second light source, and the camera. The multiple image processor causes the first light source and the second light source to turn on, such as in sequence, and also causes the camera to generate two or more sets of image data, such as one set when each of the light sources is illuminated, through the use of filters or tuned pixels, or otherwise.
- The present invention provides many important technical advantages. One important technical advantage of the present invention is a system and method for multiple image analysis that uses two or more sets of image data to analyze a component. Each set of image data is obtained when the component is illuminated by a light source having a different lighting angle, which creates shaded areas that can be analyzed to determine whether they indicate the existence of damage or unacceptable dimensional variations on the component.
- Those skilled in the art will further appreciate the advantages and superior features of the invention together with other important aspects thereof on reading the detailed description that follows in conjunction with the drawings.
- FIG. 1 is a diagram of a system for performing multiple image analysis in accordance with an exemplary embodiment of the present invention;
- FIGS. 2A, 2B, and2C show an exemplary undamaged element and corresponding bright and shaded regions generated by illumination from light sources;
- FIGS. 3A, 3B, and3C show an exemplary damaged element and corresponding bright and shaded regions generated by illumination from light sources;
- FIG. 4 is a diagram of a system for processing image data from multiple images in accordance with an exemplary embodiment of the present invention;
- FIG. 5 is a flowchart of a method for analyzing image data from multiple images in accordance with an exemplary embodiment of the present invention;
- FIG. 6 is a flowchart of a method for analyzing image data in accordance with an exemplary embodiment of the present invention; and
- FIG. 7 is a flowchart of a method for performing image data analysis for multiple images in accordance with an exemplary embodiment of the present invention.
- In the description that follows, like parts are marked throughout the specification and drawings with the same reference numerals, respectively. The drawing figures might not be to scale, and certain components can be shown in generalized or schematic form and identified by commercial designations in the interest of clarity and conciseness.
- FIG. 1 is a diagram of a system100 for performing multiple image analysis in accordance with an exemplary embodiment of the present invention. System 100 allows three-dimensional aspects of an inspected device or component to be determined from images obtained from two or more different viewing angles.
- System100 includes
multiple image processor 102, which can be implemented in hardware, software, or a suitable combination of hardware and software, and which can be one or more software systems operating on a general purpose processing platform. As used herein, a software system can include one or more objects, agents, subroutines, lines of code, threads, two or more lines of code or other suitable software structures operating in two or more separate software applications, or other suitable software structures, and can operate on two or more different processors, or other suitable configurations of processors. In one exemplary embodiment, a software system can include one or more lines of code or other software structures operating in a general purpose software application, such as an operating system, and one or more lines of code or other suitable software structures operating in a specific purpose software application. -
Multiple image processor 102 is coupled tolight sources -
Multiple image processor 102 is also coupled tocamera 106.Camera 106 can be a charge coupled device (CCD), a CMOS imaging device, or other suitable imaging devices that are focused on acomponent 108 having a plurality ofelements 110.Light sources component 108, andilluminate component 108 from different angles as shown in FIG. 1. Thus, the lightilluminating component 108 fromlight source 104 a will create shaded regions that are different from the shaded regions created by lightilluminating component 108 fromlight source 104 b. Additional light sources can be used where suitable to create additional shaded regions. - Camera106 is used to record image data of
component 108 while it is being illuminated bylight sources camera 106 is controlled bymultiple image processor 102 to store a first set of image data ofcomponent 108 whenlight source 104 a is on, and to store a second set of image data whenlight source 104 b is on. Likewise,camera 106 can store image data when both oflight sources camera 106 can record image data according to the frequency of the light that creates the image, such as by including one or more light filters, two or more sets of pixels that are tuned to received predetermined frequencies of light, or to otherwise differentiate between light illuminated fromlight sources - In operation, a
component 108 is placed in the focal area oflight sources camera 106 for inspection.Multiple image processor 102 then causescomponent 108 to be illuminated and causescamera 106 to produce image data, such as by generating an N×M array of pixels of image data, which can then be stored bymultiple image processor 102 or other suitable storage systems or devices. Because of the angular difference betweenlight sources component 108, shaded regions are generated fromelements 110.Multiple image processor 102 can analyze these shaded regions to determine whether they are indicative of any damage or defects tocomponent 108,elements 110, or other suitable indications. - In this manner,
multiple image processor 102 can determine whether three-dimensional defects or other variations incomponent 108 orelements 110 exist. For example, if one ofelements 110 is damaged, then the shaded regions generated by thatelement 110 when it is illuminated bylight sources light sources - FIGS. 2A, 2B, and2C show an exemplary
undamaged element 110 and corresponding bright and shaded regions generated by illumination fromlight sources - FIG. 2A shows an exemplary
undamaged element 110, which is semi-spherical in shape. In FIG. 2B, the circular outline ofelement 110 as viewed from overhead is shown with an illuminated region and a shaded region corresponding to the shadow generated bylight source 104 a. As shown, the shaded region generates a distinctive pattern which is indicative of a spherical configuration ofelement 110. Likewise, in FIG. 2C, the shaded region ofelement 110 is on the opposite face, as a result of the location oflight source 104 b. Thus, the shaded regions generated shown in FIGS. 2b and 2 c can be used as a reference for anundamaged element 110. - In addition, the differences in pixel brightness data between FIG. 2B and FIG. 2C and the known angle of illumination from
light sources element 110. For example, it can be determined from areas in FIG. 2B and FIG. 2C in which the pixel brightness data is a maximum and does not vary that such areas are not directly blocked from direct exposure by either source. Likewise, as the difference in brightness data increases for a given pixel of FIG. 2B and FIG. 2C, it can be determined that an obstruction is blocking those pixels, and that the obstruction is located between light source having the lower brightness values and the location of the pixel being analyzed. Other suitable procedures can be used to estimate the size and location of dimensional variations based upon pixel data, such as the use of empirically developed pass/fail ratios based upon the size of areas in which pixel brightness variations between two or more images exceed predetermined levels. - FIGS. 3A, 3B, and3C show an exemplary damaged
element 110 and corresponding bright and shaded regions generated by illumination fromlight sources - FIG. 3A shows the damaged
element 110 which varies from semi-circular, such as by an indentation. As shown in FIG. 3B, this indentation creates shadedregions shaded regions shaded region 202 in FIG. 2b. These exemplary variations can be used to detect three-dimensional variations inelement 110 that would otherwise be difficult to detect from a single image, depending on the angle of illumination. Likewise, FIG. 3C includes shadedregion 306 which varies from shadedregion 204. The pixels defining these regions can be compared between a test image, such as that shown in FIG. 3B and FIG. 3C, and a reference image, such as that shown in FIG. 2B and FIG. 2C, to determine whether the region defined by such pixel variations exceeds predetermined allowable areas for defects. Likewise, the composite images formed by combining image data fromshaded regions - In operation, the shaded regions generated by illumination from
light sources - FIG. 4 is a diagram of a system400 for processing image data from multiple images in accordance with an exemplary embodiment of the present invention. System 400 includes
multiple image processor 102 andlight sequence controller 402,first image analyzer 404,second image analyzer 406,image comparator 3D image constructor 410, each of which can be implemented in hardware, software, or a suitable combination of hardware and software, and which can be one or more software systems operating on a general purpose processor platform. -
Light sequence controller 402 controls the sequence in whichlight sources component 108. Likewise,light sequence controller 402 also controls the operation ofcamera 106, such that when a first light source is illuminating thecomponent 108,camera 106 captures first image data, and when a second light source is illuminating thecomponent 108,camera 106 captures or generates second image data. Likewise,light sequence controller 402 can control light sources having different frequencies, such thatcamera 106 can generate multiple sets of image data simultaneously so as to decrease the amount of time required to generate the multiple sets of image data. -
First image analyzer 404 andsecond image analyzer 406 receive an N×M array of pixels of brightness data, and analyze the pixel data to determine whether the pixel data is acceptable, requires additional analysis such as comparison with a reference image or dimensional analysis, or is unacceptable.First image analyzer 404 andsecond image analyzer 406 then generate status data indicating whether the pixel data is acceptable, requires further analysis, or is unacceptable. In one exemplary embodiment,first image analyzer 404 receives pixel array data generated whenlight source 104 ailluminates component 108 andsecond image analyzer 406 receives pixel array data generated whenlight source 104 b illuminatescomponent 108. Additional image analyzers can also be used to accommodate light sources illuminating thecomponent 108 from different angles.First image analyzer 404 andsecond image analyzer 406 perform pixel brightness analysis of the corresponding images. - In one exemplary embodiment,
first image analyzer 404 andsecond image analyzer 406 determine whether the pixel data indicates that the number and magnitude of variations in pixel brightness data exceed predetermined maximum allowable numbers and magnitudes, such that it is determinable whether the component contains unacceptable dimensional variations without additional image data analysis. In addition,first image analyzer 404 andsecond image analyzer 406 can determine whether the pixel data falls within a range of values that indicates that further analysis is required. -
Image comparator 408 receives first image data and second image data and generates difference image data, such as by subtracting pixel brightness data for corresponding pixels between a first image and a second image.Image comparator 408 can perform comparator analysis of first test image data and first reference image data, second test image data and second reference image data, composite test image data and composite reference image data, or other suitable sets of corresponding image data.Image comparator 408 can also generate absolute brightness variation data, relative brightness variation data, or other suitable brightness variation data. -
3D image constructor 410 can receive the test image data, reference image data, difference image data, composite image data, or other suitable image data and determine whether defects, variations, or other features ofelement 110 or other elements exceed allowable variations for such elements. In one exemplary embodiment,3D image constructor 410 can determine from the known angle of illumination oflight sources element 110 at that corresponding position. Likewise,3D image constructor 410 can include predetermined ranges for allowable variations, such as histogram data, pixel area mapping data, and other suitable data. In this manner,3D image constructor 410 can be used to generate dimensional variation data after determining whether a variation or feature in anelement 110 exceeds allowable limits, such that the component having the element can be rejected in the event the damage or dimensional variation in theelement 110 exceeds such limits. - In operation, system400 is used to control the inspection of a component, to generate test image data, to analyze the test image data, and to estimate three-dimensional variations or features of a test image. System 400 utilizes image data generated by illuminating the component from two or more angles, can combine the test image data and compare the test image data to reference image data, and can process any difference image data to make determinations on whether or not to accept or reject a component.
- FIG. 5 is a flowchart of a method500 for analyzing image data from multiple images in accordance with an exemplary embodiment of the present invention. Method 500 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
- Method500 begins at 502 where image data is obtained. In one exemplary embodiment, the image data is obtained by simultaneously illuminating a component with multiple light sources from different angles, where each light source is illuminated at a different time. Likewise, the light sources can provide light having different frequencies where the image data is generated at the same time and filters, tuned pixels, or other procedures are used to separate the image data created from each light source. The method then proceeds to 504.
- At504 each set of image data is analyzed. The sets of image data can be analyzed by generating histogram data showing the brightness of each pixel, by comparing each set of test image data to a set of reference image data and performing histogram analysis or other suitable analysis of the difference image data set, by combining the test image data and comparing the combined test image data to predetermined acceptable ranges for histogram data, by comparing the combined test image data to combined reference image data, or performing other suitable analyses. The method then proceeds to 506.
- At506 it is determined whether all images are within an acceptable predetermined range. If all images are within such range, the method proceeds to 508 where the image is accepted and any subsequent analysis is performed on that component, other components can be selected for analysis, or other suitable procedures can be implemented. Otherwise the method proceeds to 510.
- At510 compared image data is obtained. In one exemplary embodiment, an initial analysis is performed at 504 to determine whether additional analyses need to be performed, such that the analysis performed at 504 does not include comparator data. Other suitable processes can be used. After the comparator image data is obtained for each single test image and reference image, composite test and reference images, or other suitable comparative data, the method proceeds to 512.
- At512 the image data and comparator data is analyzed to generate three-dimensional image data. In one exemplary embodiment, the three-dimensional image data can include predetermined allowable ranges for three-dimensional variations that generate shaded regions of elements when illuminated by multiple light sources. Likewise, the three-dimensional image data can include estimates of variations and components based upon the known angular relationship between the light sources and the component. The method then proceeds to 514.
- At514 the three-dimensional image data is applied to template data. In one exemplary embodiment, the template data can include one or more templates that are used to estimate variations between measured brightness data and expected brightness data, so as to determine whether three-dimensional variations in the inspected component exceed allowable variations. The method then proceeds to 516.
- At516 it is determined whether the three-dimensional data is within a predetermined range. If the three-dimensional data is not within the range, the method proceeds to 518 where the image data is rejected, such as by rejecting the component, flagging the component for further manual inspection, or other suitable procedures. Otherwise, the method proceeds to 520 where the image data is accepted.
- In operation, method500 is used to analyze multiple sets of image data for a test component in order to determine whether the component includes dimensional variations, damage, or other unacceptable condition. Method 500 further utilizes light sources having different angular relationships to the test component, where the known angular relationship of the light sources can be used in conjunction with the pixel brightness data to estimate 3-dimensional variations in the test component.
- FIG. 6 is a flowchart of a method600 for analyzing image data in accordance with an exemplary embodiment of the present invention. Method 600 can be used to perform component image analysis to detect damaged components, or for other suitable purposes.
- Method600 begins at 602 where a test piece is exposed to light from two different light frequencies and two different angular illumination zones. The method then proceeds to 604 and 608 in parallel. At 604, first image data is obtained, such as by filtering the light through a first filter, by using pixels tuned to the first light frequency, or other suitable methods. Likewise, at 608, the second image data is obtained, such as by filtering the light through a second filter, by using pixels tuned to a second light frequency, or other suitable methods. The method then proceeds to 606 from 604 and 610 from 608, respectively.
- At606, the pixel brightness variation data is analyzed for the first image data. For example, pixel histogram data can be generated and the variations in pixel brightness can be compared to predetermined acceptable ranges. Likewise, other suitable pixel brightness variation analysis methods can be used. At 610, similar pixel brightness variations are analyzed for the second image data. The method then proceeds to 612 and 614, respectively.
- At612 and 614 it is determined whether the variations in pixel brightness for the first image data and the second image data are within a predetermined range. If both sets of image data have acceptable variations then the method proceeds to 618 and the image data is accepted. Likewise, if either of the images has unacceptable range variations the method proceeds to 616 where three-dimensional analysis is performed to determine whether the dimensional variations in the test piece are acceptable.
- In operation, method600 can be used to determine whether three-dimensional analysis of component image data should be performed, such as to perform a quick preliminary component image inspection analysis for the purpose of determining whether additional analysis should be performed. Method 600 also allows image data to be analyzed in parallel where suitable, such as when a parallel processor platform is being used to analyze the image data. Alternatively, method 600 can be implemented on a general purpose processing platform, such as through the use of multitasking or by otherwise simulating parallel processing.
- FIG. 7 is a flowchart of a
method 700 for performing image data analysis for multiple images in accordance with an exemplary embodiment of the present invention.Method 700 can be used to perform component image analysis to detect damaged components, or for other suitable purposes. -
Method 700 begins at 702 and 704 in parallel. At 702, first reference image data is compared to first test image data, and at 704, second reference image data is compared to second test image data. This comparison can include a pixel to correlating pixel brightness subtraction to generate a difference image, or other suitable comparison procedures. The method then proceeds to 706. - At706 it is determined whether acceptable variations exist in the compare data, such as by generating a histogram having pixel frequency and magnitude for difference data. If it is determined that the variations are acceptable the method proceeds to 708 where the image data is accepted. Likewise, if the variations are not acceptable the method proceeds to 710.
- At710 a composite test image is formed. In one exemplary embodiment, the composite test image can include two or more sets of image data generated from two or more different illumination angles, from two or more different light frequencies, or other suitable composite test data. The method then proceeds to 712.
- At712 the composite test image data is compared to composite reference image data, such as by performing a pixel to corresponding pixel subtraction or other suitable compare procedures. The method then proceeds to 714.
- At714, three-dimensional coordinates for the component being inspected are estimated from variations in the test image data as compared to the reference image data. For example, pixels at coordinates that have significant variations in brightness as a function of the angle of illumination can indicate the existence of an indentation, spur, bulge, or other deformity in the component. It may be determined by analysis, empirically, or otherwise that such variations in brightness that exceed certain levels correlate to dimensional variations. Likewise, an estimate of the dimensional variation can be calculated from the brightness data and the known angular position of each light source. Other suitable methods may also or alternatively be used. The method then proceeds to 716.
- At716, it is determined whether the variations in the component dimensions are allowable. In one exemplary embodiment, allowable variations can be determined empirically, by calculation, can be set according to a customer or industry standard, or through other suitable methods. The method then proceeds to 718 if it is determined that the variations exceed allowable ranges and the image data is rejected. Likewise, a message can be generated informing the operator that additional analysis or operator inspection is required. If it is determined at 716 that any variations in the image data are allowable then the method proceeds to 720 where the image data is accepted.
- In operation,
method 700 allows a component to be inspected by illuminating the component from multiple light sources, such that the component generates shaded regions and bright regions. The shaded and bright regions of the component can be then analyzed and compared to reference image data to determine whether unacceptable variations or damage may exist on the component. - Although exemplary embodiments of a system and method for multiple image analysis have been described in detail herein, those skilled in the art will also recognize that various substitutions and modifications can be made to the systems and methods without departing from the scope and spirit of the appended claims.
Claims (20)
1. A system for multiple image analysis comprising:
a first light source;
a second light source;
a camera; and
a multiple image processor coupled to the first light source, the second light source, and the camera, the multiple image processor causing the first light source and the second light source to turn on and the camera to generate two or more sets of image data.
2. The system of claim 1 wherein the first light source emits light having a first frequency and the second light source emits light having a second frequency.
3. The system of claim 2 wherein the camera can generate two or more sets of image data when both the first light source and the second light source are emitting light.
4. The system of claim 2 wherein the camera further comprises:
a first set of pixels receiving light at the first frequency; and
a second set of pixels receiving light at the second frequency.
5. The system of claim 2 wherein the camera further comprises:
a first filter passing light at the first frequency; and
a second filter passing light at the second frequency.
6. The system of claim 1 wherein the multiple image processor further comprises a light sequence controller causing the first light source and the second light source to turn on and turn off.
7. The system of claim 1 wherein the multiple image processor further comprises an image analyzer receiving the two or more sets of image data and generating status data that indicates whether the image data is acceptable.
8. The system of claim 1 wherein the multiple image processor further comprises a first image analyzer receiving the first set of image data and a second image analyzer receiving the second set of image data and generating status data that indicates whether the image data is acceptable.
9. The system of claim 1 wherein the multiple image processor further comprises an image comparator receiving the two or more sets of image data and generating difference data.
10. The system of claim 1 wherein the multiple image processor further comprises an image constructor receiving the two or more sets of image data and generating dimensional variation data.
11. A method for inspecting a component comprising:
illuminating the component from a first illumination angle;
receiving first image data of the component;
illuminating the component from a second illumination angle;
receiving second image data of the component; and
using the first image data and the second image data to determine whether a dimension of the component is acceptable.
12. The method of claim 11 wherein illuminating the component from the first illumination angle and illuminating the component from the second illumination angle further comprises illuminating the component using light having a first frequency from the first illumination angle and illuminating the component using light having a second frequency from the second illumination angle.
13. The method of claim 11 wherein receiving the first image data of the component comprises receiving the first image data of the component by filtering light received from the component.
14. The method of claim 11 wherein receiving the first image data of the component and receiving the second image data of the component comprises receiving the first image data of the component by filtering light received from the component with a first filter and receiving the second image data of the component by filtering light received from the component with a second filter.
15. The method of claim 11 wherein receiving the first image data of the component comprises receiving the first image data of the component with a first set of pixels.
16. The method of claim 11 wherein receiving the first image data of the component and receiving the second image data of the component comprises receiving the first image data of the component with a first set of pixels and receiving the second image data of the component with a second set of pixels.
17. A method for inspecting a component comprising:
receiving first image data and second image data of the component;
comparing the first image data to reference image data to generate first difference data;
comparing the second image data to reference image data to generate second difference data; and
generating component dimension data from the first difference data and the second difference data.
18. The method of claim 17 further comprising:
combining the first image data and the second image data to generate composite image data;
comparing the composite image data to composite reference data to generate composite difference data; and
generating component dimension data from the composite difference data.
19. The method of claim 17 wherein the step of receiving the first image data and the second image data of the component is preceded by the step of receiving status data that indicates that the component requires additional analysis to determine whether it has unacceptable dimensional variations.
20. The method of claim 17 wherein generating the component dimension data from the first difference data and the second difference data further comprises using light source angular data to generate the component dimension data.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/876,795 US20020186878A1 (en) | 2001-06-07 | 2001-06-07 | System and method for multiple image analysis |
PCT/IB2002/002050 WO2002099403A1 (en) | 2001-06-07 | 2002-06-05 | System and method for multiple image analysis |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/876,795 US20020186878A1 (en) | 2001-06-07 | 2001-06-07 | System and method for multiple image analysis |
Publications (1)
Publication Number | Publication Date |
---|---|
US20020186878A1 true US20020186878A1 (en) | 2002-12-12 |
Family
ID=25368601
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US09/876,795 Abandoned US20020186878A1 (en) | 2001-06-07 | 2001-06-07 | System and method for multiple image analysis |
Country Status (2)
Country | Link |
---|---|
US (1) | US20020186878A1 (en) |
WO (1) | WO2002099403A1 (en) |
Cited By (81)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030169917A1 (en) * | 2002-01-15 | 2003-09-11 | Nec Corporation | Three-dimensional body photography apparatus, three-dimensional geometricmodel generation apparatus, three-dimensional geometric model generation method, and record medium in which three-dimensional geometric model generation program was stored |
US20040168881A1 (en) * | 2003-01-10 | 2004-09-02 | National Rejectors Inc. Gmbh | Process for identifying an embossed image of a coin in an automatice coin machine |
US20040245334A1 (en) * | 2003-06-06 | 2004-12-09 | Sikorski Steven Maurice | Inverted terminal presentation scanner and holder |
US20050103753A1 (en) * | 2003-11-18 | 2005-05-19 | Matsushita Electric Industrial Co., Ltd. | Light processing apparatus |
US20050286754A1 (en) * | 2004-06-14 | 2005-12-29 | Thomas Steven P | Electronic component detection system |
US20060006311A1 (en) * | 2001-06-22 | 2006-01-12 | Orbotech Ltd | High-sensitivity optical scanning using memory integration |
US20060291714A1 (en) * | 2004-12-07 | 2006-12-28 | Kenong Wu | Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle |
GB2427541A (en) * | 2005-06-24 | 2007-01-03 | Aew Delford Group Ltd | Vision system for food cutting and portioning apparatus |
GB2428182A (en) * | 2005-06-24 | 2007-01-24 | Aew Delford Systems | Vision system for food cutting and portioning apparatus |
US20070027585A1 (en) * | 2005-07-27 | 2007-02-01 | Thomas Wulff | System and method for monitoring a mobile computing product/arrangement |
US20070076943A1 (en) * | 2005-04-15 | 2007-04-05 | Vistec Semiconductor Systems Jena Gmbh | Method and apparatus for inspecting a wafer |
US20070297028A1 (en) * | 2006-06-21 | 2007-12-27 | Thomas Wulff | System and device for monitoring a computing device |
US20070298751A1 (en) * | 2006-06-21 | 2007-12-27 | Thomas Wulff | System and method for monitoring a mobile device |
US7646906B2 (en) | 2004-01-29 | 2010-01-12 | Kla-Tencor Technologies Corp. | Computer-implemented methods for detecting defects in reticle design data |
US7676077B2 (en) | 2005-11-18 | 2010-03-09 | Kla-Tencor Technologies Corp. | Methods and systems for utilizing design data in combination with inspection data |
US7689966B2 (en) | 2004-09-14 | 2010-03-30 | Kla-Tencor Technologies Corp. | Methods, systems, and carrier media for evaluating reticle layout data |
US20100100623A1 (en) * | 2004-04-06 | 2010-04-22 | Thomas Wulff | System and method for monitoring a mobile computing product/arrangement |
US7711514B2 (en) | 2007-08-10 | 2010-05-04 | Kla-Tencor Technologies Corp. | Computer-implemented methods, carrier media, and systems for generating a metrology sampling plan |
US7738093B2 (en) | 2007-05-07 | 2010-06-15 | Kla-Tencor Corp. | Methods for detecting and classifying defects on a reticle |
US7769225B2 (en) * | 2005-08-02 | 2010-08-03 | Kla-Tencor Technologies Corp. | Methods and systems for detecting defects in a reticle design pattern |
US7796804B2 (en) | 2007-07-20 | 2010-09-14 | Kla-Tencor Corp. | Methods for generating a standard reference die for use in a die to standard reference die inspection and methods for inspecting a wafer |
US20100245813A1 (en) * | 2009-03-24 | 2010-09-30 | Orbotech Ltd. | Multi-modal imaging |
US7877722B2 (en) | 2006-12-19 | 2011-01-25 | Kla-Tencor Corp. | Systems and methods for creating inspection recipes |
US7962863B2 (en) | 2007-05-07 | 2011-06-14 | Kla-Tencor Corp. | Computer-implemented methods, systems, and computer-readable media for determining a model for predicting printability of reticle features on a wafer |
US7975245B2 (en) | 2007-08-20 | 2011-07-05 | Kla-Tencor Corp. | Computer-implemented methods for determining if actual defects are potentially systematic defects or potentially random defects |
US20110206237A1 (en) * | 2010-02-25 | 2011-08-25 | Canon Kabushiki Kaisha | Recognition apparatus and method thereof, and computer program |
US8041103B2 (en) | 2005-11-18 | 2011-10-18 | Kla-Tencor Technologies Corp. | Methods and systems for determining a position of inspection data in design data space |
US8112241B2 (en) | 2009-03-13 | 2012-02-07 | Kla-Tencor Corp. | Methods and systems for generating an inspection process for a wafer |
US20120031975A1 (en) * | 2010-08-04 | 2012-02-09 | The Code Corporation | Illumination blocks for a graphical code reader |
US8139844B2 (en) | 2008-04-14 | 2012-03-20 | Kla-Tencor Corp. | Methods and systems for determining a defect criticality index for defects on wafers |
US8194968B2 (en) | 2007-01-05 | 2012-06-05 | Kla-Tencor Corp. | Methods and systems for using electrical information for a device being fabricated on a wafer to perform one or more defect-related functions |
US8204297B1 (en) | 2009-02-27 | 2012-06-19 | Kla-Tencor Corp. | Methods and systems for classifying defects detected on a reticle |
US8213704B2 (en) | 2007-05-09 | 2012-07-03 | Kla-Tencor Corp. | Methods and systems for detecting defects in a reticle design pattern |
US20130144797A1 (en) * | 2008-10-02 | 2013-06-06 | ecoATM, Inc. | Method And Apparatus For Recycling Electronic Devices |
CN103674974A (en) * | 2012-09-06 | 2014-03-26 | 株式会社岛津制作所 | Inspection apparatus |
US8775101B2 (en) | 2009-02-13 | 2014-07-08 | Kla-Tencor Corp. | Detecting defects on a wafer |
US8781781B2 (en) | 2010-07-30 | 2014-07-15 | Kla-Tencor Corp. | Dynamic care areas |
US8826200B2 (en) | 2012-05-25 | 2014-09-02 | Kla-Tencor Corp. | Alteration for wafer inspection |
US8831334B2 (en) | 2012-01-20 | 2014-09-09 | Kla-Tencor Corp. | Segmentation for wafer inspection |
US8923600B2 (en) | 2005-11-18 | 2014-12-30 | Kla-Tencor Technologies Corp. | Methods and systems for utilizing design data in combination with inspection data |
US9053527B2 (en) | 2013-01-02 | 2015-06-09 | Kla-Tencor Corp. | Detecting defects on a wafer |
US9087367B2 (en) | 2011-09-13 | 2015-07-21 | Kla-Tencor Corp. | Determining design coordinates for wafer defects |
US9092846B2 (en) | 2013-02-01 | 2015-07-28 | Kla-Tencor Corp. | Detecting defects on a wafer using defect-specific and multi-channel information |
US9134254B2 (en) | 2013-01-07 | 2015-09-15 | Kla-Tencor Corp. | Determining a position of inspection system output in design data space |
US9170211B2 (en) | 2011-03-25 | 2015-10-27 | Kla-Tencor Corp. | Design-based inspection using repeating structures |
US9189844B2 (en) | 2012-10-15 | 2015-11-17 | Kla-Tencor Corp. | Detecting defects on a wafer using defect-specific information |
JP2016035405A (en) * | 2014-08-01 | 2016-03-17 | リコーエレメックス株式会社 | Image inspection device, image inspection system, and image inspection method |
US9311698B2 (en) | 2013-01-09 | 2016-04-12 | Kla-Tencor Corp. | Detecting defects on a wafer using template image matching |
US9310320B2 (en) | 2013-04-15 | 2016-04-12 | Kla-Tencor Corp. | Based sampling and binning for yield critical defects |
US20160267339A1 (en) * | 2015-03-11 | 2016-09-15 | Canon Kabushiki Kaisha | Image processing apparatus and method of generating face image |
JP2017032548A (en) * | 2015-06-26 | 2017-02-09 | コグネックス・コーポレイション | Using 3d vision for automated industrial inspection |
US9659670B2 (en) | 2008-07-28 | 2017-05-23 | Kla-Tencor Corp. | Computer-implemented methods, computer-readable media, and systems for classifying defects detected in a memory device area on a wafer |
JP2017122682A (en) * | 2016-01-08 | 2017-07-13 | 株式会社キーエンス | Visual inspection device, visual inspection method, and computer program executable by controller used in visual inspection device |
US9818160B2 (en) | 2008-10-02 | 2017-11-14 | ecoATM, Inc. | Kiosk for recycling electronic devices |
US9865512B2 (en) | 2013-04-08 | 2018-01-09 | Kla-Tencor Corp. | Dynamic design attributes for wafer inspection |
US9881284B2 (en) | 2008-10-02 | 2018-01-30 | ecoATM, Inc. | Mini-kiosk for recycling electronic devices |
US9885672B2 (en) | 2016-06-08 | 2018-02-06 | ecoATM, Inc. | Methods and systems for detecting screen covers on electronic devices |
US9904911B2 (en) | 2008-10-02 | 2018-02-27 | ecoATM, Inc. | Secondary market and vending system for devices |
US9911102B2 (en) | 2014-10-02 | 2018-03-06 | ecoATM, Inc. | Application for device evaluation and other processes associated with device recycling |
TWI626438B (en) * | 2015-09-30 | 2018-06-11 | 佳能股份有限公司 | Inspection apparatus and article manufacturing method |
US10032140B2 (en) * | 2008-10-02 | 2018-07-24 | ecoATM, LLC. | Systems for recycling consumer electronic devices |
US10127647B2 (en) | 2016-04-15 | 2018-11-13 | Ecoatm, Llc | Methods and systems for detecting cracks in electronic devices |
US10269110B2 (en) | 2016-06-28 | 2019-04-23 | Ecoatm, Llc | Methods and systems for detecting cracks in illuminated electronic device screens |
US10401411B2 (en) | 2014-09-29 | 2019-09-03 | Ecoatm, Llc | Maintaining sets of cable components used for wired analysis, charging, or other interaction with portable electronic devices |
US10417615B2 (en) | 2014-10-31 | 2019-09-17 | Ecoatm, Llc | Systems and methods for recycling consumer electronic devices |
US10445708B2 (en) | 2014-10-03 | 2019-10-15 | Ecoatm, Llc | System for electrically testing mobile devices at a consumer-operated kiosk, and associated devices and methods |
US10475002B2 (en) | 2014-10-02 | 2019-11-12 | Ecoatm, Llc | Wireless-enabled kiosk for recycling consumer devices |
US10572946B2 (en) | 2014-10-31 | 2020-02-25 | Ecoatm, Llc | Methods and systems for facilitating processes associated with insurance services and/or other services for electronic devices |
WO2020049551A1 (en) | 2018-09-06 | 2020-03-12 | Orbotech Ltd. | Multimodality multiplexed illumination for optical inspection systems |
US10825082B2 (en) | 2008-10-02 | 2020-11-03 | Ecoatm, Llc | Apparatus and method for recycling mobile phones |
US10860990B2 (en) | 2014-11-06 | 2020-12-08 | Ecoatm, Llc | Methods and systems for evaluating and recycling electronic devices |
US11010841B2 (en) | 2008-10-02 | 2021-05-18 | Ecoatm, Llc | Kiosk for recycling electronic devices |
US11080672B2 (en) | 2014-12-12 | 2021-08-03 | Ecoatm, Llc | Systems and methods for recycling consumer electronic devices |
US20210350495A1 (en) * | 2020-02-10 | 2021-11-11 | Cognex Corporation | Composite three-dimensional blob tool and method for operating the same |
US20220292665A1 (en) * | 2019-10-02 | 2022-09-15 | Konica Minolta, Inc. | Workpiece surface defect detection device and detection method, workpiece surface inspection system, and program |
US11462868B2 (en) | 2019-02-12 | 2022-10-04 | Ecoatm, Llc | Connector carrier for electronic device kiosk |
US11482067B2 (en) | 2019-02-12 | 2022-10-25 | Ecoatm, Llc | Kiosk for evaluating and purchasing used electronic devices |
US20230084949A1 (en) * | 2021-09-13 | 2023-03-16 | Quanta Computer Inc. | Systems and methods for monitoring antenna arrays |
US11798250B2 (en) | 2019-02-18 | 2023-10-24 | Ecoatm, Llc | Neural network based physical condition evaluation of electronic devices, and associated systems and methods |
US11922467B2 (en) | 2020-08-17 | 2024-03-05 | ecoATM, Inc. | Evaluating an electronic device using optical character recognition |
US11974046B2 (en) | 2019-08-27 | 2024-04-30 | Orbotech Ltd. | Multimodality multiplexed illumination for optical inspection systems |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5822055A (en) * | 1995-06-06 | 1998-10-13 | Kla Instruments Corporation | Optical inspection of a specimen using multi-channel responses from the specimen using bright and darkfield detection |
US5946029A (en) * | 1996-06-25 | 1999-08-31 | Matsushita Electric Works, Ltd | Image processing process |
US5982493A (en) * | 1998-06-02 | 1999-11-09 | Motorola, Inc. | Apparatus and method for acquiring multiple images |
US5995220A (en) * | 1998-02-19 | 1999-11-30 | Komatsu, Ltd. | Semiconductor package inspection apparatus |
US6075883A (en) * | 1996-11-12 | 2000-06-13 | Robotic Vision Systems, Inc. | Method and system for imaging an object or pattern |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS61293657A (en) * | 1985-06-21 | 1986-12-24 | Matsushita Electric Works Ltd | Method for inspecting soldering appearance |
JP2953736B2 (en) * | 1990-03-20 | 1999-09-27 | 松下電器産業株式会社 | Solder shape inspection method |
US5064291A (en) * | 1990-04-03 | 1991-11-12 | Hughes Aircraft Company | Method and apparatus for inspection of solder joints utilizing shape determination from shading |
JP3072998B2 (en) * | 1990-04-18 | 2000-08-07 | 株式会社日立製作所 | Soldering condition inspection method and apparatus |
DE19725633C1 (en) * | 1997-06-17 | 1998-12-17 | Zentrum Fuer Neuroinformatik G | Method and arrangement for analyzing the nature of a surface |
-
2001
- 2001-06-07 US US09/876,795 patent/US20020186878A1/en not_active Abandoned
-
2002
- 2002-06-05 WO PCT/IB2002/002050 patent/WO2002099403A1/en not_active Application Discontinuation
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5822055A (en) * | 1995-06-06 | 1998-10-13 | Kla Instruments Corporation | Optical inspection of a specimen using multi-channel responses from the specimen using bright and darkfield detection |
US5946029A (en) * | 1996-06-25 | 1999-08-31 | Matsushita Electric Works, Ltd | Image processing process |
US6075883A (en) * | 1996-11-12 | 2000-06-13 | Robotic Vision Systems, Inc. | Method and system for imaging an object or pattern |
US5995220A (en) * | 1998-02-19 | 1999-11-30 | Komatsu, Ltd. | Semiconductor package inspection apparatus |
US5982493A (en) * | 1998-06-02 | 1999-11-09 | Motorola, Inc. | Apparatus and method for acquiring multiple images |
Cited By (129)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060006311A1 (en) * | 2001-06-22 | 2006-01-12 | Orbotech Ltd | High-sensitivity optical scanning using memory integration |
US7129509B2 (en) * | 2001-06-22 | 2006-10-31 | Orbotech, Ltd. | High-sensitivity optical scanning using memory integration |
US8536506B2 (en) | 2001-06-22 | 2013-09-17 | Orbotech Ltd. | Imaging device and method for high-sensitivity optical scanning and integrated circuit therefor |
US8119969B2 (en) | 2001-06-22 | 2012-02-21 | Orbotech Ltd | Imaging device and method for high-sensitivity optical scanning and integrated circuit therefor |
US9232114B2 (en) | 2001-06-22 | 2016-01-05 | Orbotech Ltd. | Imaging device and method for high-sensitivity optical scanning and integrated circuit therefor |
US7272255B2 (en) * | 2002-01-15 | 2007-09-18 | Nec Corporation | Three-dimensional body photography apparatus, three-dimensional geometricmodel generation apparatus, three-dimensional geometric model generation method, and record medium in which three-dimensional geometric model generation program was stored |
US20030169917A1 (en) * | 2002-01-15 | 2003-09-11 | Nec Corporation | Three-dimensional body photography apparatus, three-dimensional geometricmodel generation apparatus, three-dimensional geometric model generation method, and record medium in which three-dimensional geometric model generation program was stored |
US20040168881A1 (en) * | 2003-01-10 | 2004-09-02 | National Rejectors Inc. Gmbh | Process for identifying an embossed image of a coin in an automatice coin machine |
US7004299B2 (en) * | 2003-01-10 | 2006-02-28 | National Rejectors, Inc. Gmbh | Process for identifying an embossed image of a coin in an automatic coin machine |
US20040245334A1 (en) * | 2003-06-06 | 2004-12-09 | Sikorski Steven Maurice | Inverted terminal presentation scanner and holder |
US20050103753A1 (en) * | 2003-11-18 | 2005-05-19 | Matsushita Electric Industrial Co., Ltd. | Light processing apparatus |
US7532749B2 (en) * | 2003-11-18 | 2009-05-12 | Panasonic Corporation | Light processing apparatus |
US7646906B2 (en) | 2004-01-29 | 2010-01-12 | Kla-Tencor Technologies Corp. | Computer-implemented methods for detecting defects in reticle design data |
US20110221673A1 (en) * | 2004-04-06 | 2011-09-15 | Symbol Technologies, Inc. | System and method for monitoring a mobile computing product/arrangement |
US8773260B2 (en) | 2004-04-06 | 2014-07-08 | Symbol Technologies, Inc. | System and method for monitoring a mobile computing product/arrangement |
US20110205376A1 (en) * | 2004-04-06 | 2011-08-25 | Symbol Technologies, Inc. | System and method for monitoring a mobile computing product/arrangement |
US20100100623A1 (en) * | 2004-04-06 | 2010-04-22 | Thomas Wulff | System and method for monitoring a mobile computing product/arrangement |
US20110205076A1 (en) * | 2004-04-06 | 2011-08-25 | Symbol Technologies, Inc. | System and method for monitoring a mobile compputing product/arrangement |
US7551765B2 (en) * | 2004-06-14 | 2009-06-23 | Delphi Technologies, Inc. | Electronic component detection system |
US20050286754A1 (en) * | 2004-06-14 | 2005-12-29 | Thomas Steven P | Electronic component detection system |
US7689966B2 (en) | 2004-09-14 | 2010-03-30 | Kla-Tencor Technologies Corp. | Methods, systems, and carrier media for evaluating reticle layout data |
US8111900B2 (en) | 2004-12-07 | 2012-02-07 | Kla-Tencor Technologies Corp. | Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle |
US20060291714A1 (en) * | 2004-12-07 | 2006-12-28 | Kenong Wu | Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle |
US20100226562A1 (en) * | 2004-12-07 | 2010-09-09 | Kla-Tencor Technologies Corporation | Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle |
US7729529B2 (en) * | 2004-12-07 | 2010-06-01 | Kla-Tencor Technologies Corp. | Computer-implemented methods for detecting and/or sorting defects in a design pattern of a reticle |
US20070076943A1 (en) * | 2005-04-15 | 2007-04-05 | Vistec Semiconductor Systems Jena Gmbh | Method and apparatus for inspecting a wafer |
GB2427541A (en) * | 2005-06-24 | 2007-01-03 | Aew Delford Group Ltd | Vision system for food cutting and portioning apparatus |
GB2428182A (en) * | 2005-06-24 | 2007-01-24 | Aew Delford Systems | Vision system for food cutting and portioning apparatus |
US7822513B2 (en) | 2005-07-27 | 2010-10-26 | Symbol Technologies, Inc. | System and method for monitoring a mobile computing product/arrangement |
US20070027585A1 (en) * | 2005-07-27 | 2007-02-01 | Thomas Wulff | System and method for monitoring a mobile computing product/arrangement |
US7769225B2 (en) * | 2005-08-02 | 2010-08-03 | Kla-Tencor Technologies Corp. | Methods and systems for detecting defects in a reticle design pattern |
US7676077B2 (en) | 2005-11-18 | 2010-03-09 | Kla-Tencor Technologies Corp. | Methods and systems for utilizing design data in combination with inspection data |
US8139843B2 (en) | 2005-11-18 | 2012-03-20 | Kla-Tencor Technologies Corp. | Methods and systems for utilizing design data in combination with inspection data |
US8923600B2 (en) | 2005-11-18 | 2014-12-30 | Kla-Tencor Technologies Corp. | Methods and systems for utilizing design data in combination with inspection data |
US8041103B2 (en) | 2005-11-18 | 2011-10-18 | Kla-Tencor Technologies Corp. | Methods and systems for determining a position of inspection data in design data space |
US20070298751A1 (en) * | 2006-06-21 | 2007-12-27 | Thomas Wulff | System and method for monitoring a mobile device |
US8594742B2 (en) | 2006-06-21 | 2013-11-26 | Symbol Technologies, Inc. | System and method for monitoring a mobile device |
US20070297028A1 (en) * | 2006-06-21 | 2007-12-27 | Thomas Wulff | System and device for monitoring a computing device |
US7877722B2 (en) | 2006-12-19 | 2011-01-25 | Kla-Tencor Corp. | Systems and methods for creating inspection recipes |
US8194968B2 (en) | 2007-01-05 | 2012-06-05 | Kla-Tencor Corp. | Methods and systems for using electrical information for a device being fabricated on a wafer to perform one or more defect-related functions |
US7962863B2 (en) | 2007-05-07 | 2011-06-14 | Kla-Tencor Corp. | Computer-implemented methods, systems, and computer-readable media for determining a model for predicting printability of reticle features on a wafer |
US7738093B2 (en) | 2007-05-07 | 2010-06-15 | Kla-Tencor Corp. | Methods for detecting and classifying defects on a reticle |
US8213704B2 (en) | 2007-05-09 | 2012-07-03 | Kla-Tencor Corp. | Methods and systems for detecting defects in a reticle design pattern |
US8204296B2 (en) | 2007-07-20 | 2012-06-19 | Kla-Tencor Corp. | Methods for generating a standard reference die for use in a die to standard reference die inspection and methods for inspecting a wafer |
US7796804B2 (en) | 2007-07-20 | 2010-09-14 | Kla-Tencor Corp. | Methods for generating a standard reference die for use in a die to standard reference die inspection and methods for inspecting a wafer |
US7711514B2 (en) | 2007-08-10 | 2010-05-04 | Kla-Tencor Technologies Corp. | Computer-implemented methods, carrier media, and systems for generating a metrology sampling plan |
US7975245B2 (en) | 2007-08-20 | 2011-07-05 | Kla-Tencor Corp. | Computer-implemented methods for determining if actual defects are potentially systematic defects or potentially random defects |
US8139844B2 (en) | 2008-04-14 | 2012-03-20 | Kla-Tencor Corp. | Methods and systems for determining a defect criticality index for defects on wafers |
US9659670B2 (en) | 2008-07-28 | 2017-05-23 | Kla-Tencor Corp. | Computer-implemented methods, computer-readable media, and systems for classifying defects detected in a memory device area on a wafer |
US10853873B2 (en) | 2008-10-02 | 2020-12-01 | Ecoatm, Llc | Kiosks for evaluating and purchasing used electronic devices and related technology |
US11935138B2 (en) | 2008-10-02 | 2024-03-19 | ecoATM, Inc. | Kiosk for recycling electronic devices |
US9881284B2 (en) | 2008-10-02 | 2018-01-30 | ecoATM, Inc. | Mini-kiosk for recycling electronic devices |
US10055798B2 (en) | 2008-10-02 | 2018-08-21 | Ecoatm, Llc | Kiosk for recycling electronic devices |
US11107046B2 (en) | 2008-10-02 | 2021-08-31 | Ecoatm, Llc | Secondary market and vending system for devices |
US9818160B2 (en) | 2008-10-02 | 2017-11-14 | ecoATM, Inc. | Kiosk for recycling electronic devices |
US11790328B2 (en) | 2008-10-02 | 2023-10-17 | Ecoatm, Llc | Secondary market and vending system for devices |
US11907915B2 (en) | 2008-10-02 | 2024-02-20 | Ecoatm, Llc | Secondary market and vending system for devices |
US11443289B2 (en) | 2008-10-02 | 2022-09-13 | Ecoatm, Llc | Secondary market and vending system for devices |
US10032140B2 (en) * | 2008-10-02 | 2018-07-24 | ecoATM, LLC. | Systems for recycling consumer electronic devices |
US11526932B2 (en) | 2008-10-02 | 2022-12-13 | Ecoatm, Llc | Kiosks for evaluating and purchasing used electronic devices and related technology |
US10825082B2 (en) | 2008-10-02 | 2020-11-03 | Ecoatm, Llc | Apparatus and method for recycling mobile phones |
US11080662B2 (en) | 2008-10-02 | 2021-08-03 | Ecoatm, Llc | Secondary market and vending system for devices |
US10157427B2 (en) | 2008-10-02 | 2018-12-18 | Ecoatm, Llc | Kiosk for recycling electronic devices |
US20130144797A1 (en) * | 2008-10-02 | 2013-06-06 | ecoATM, Inc. | Method And Apparatus For Recycling Electronic Devices |
US11010841B2 (en) | 2008-10-02 | 2021-05-18 | Ecoatm, Llc | Kiosk for recycling electronic devices |
US9904911B2 (en) | 2008-10-02 | 2018-02-27 | ecoATM, Inc. | Secondary market and vending system for devices |
US8775101B2 (en) | 2009-02-13 | 2014-07-08 | Kla-Tencor Corp. | Detecting defects on a wafer |
US8204297B1 (en) | 2009-02-27 | 2012-06-19 | Kla-Tencor Corp. | Methods and systems for classifying defects detected on a reticle |
US8112241B2 (en) | 2009-03-13 | 2012-02-07 | Kla-Tencor Corp. | Methods and systems for generating an inspection process for a wafer |
US20100245813A1 (en) * | 2009-03-24 | 2010-09-30 | Orbotech Ltd. | Multi-modal imaging |
US8144973B2 (en) | 2009-03-24 | 2012-03-27 | Orbotech Ltd. | Multi-modal imaging |
JP2011175600A (en) * | 2010-02-25 | 2011-09-08 | Canon Inc | Recognition apparatus and method thereof, and computer program |
US9087265B2 (en) * | 2010-02-25 | 2015-07-21 | Canon Kabushiki Kaisha | Recognition apparatus and method thereof, and computer program |
US20110206237A1 (en) * | 2010-02-25 | 2011-08-25 | Canon Kabushiki Kaisha | Recognition apparatus and method thereof, and computer program |
US8781781B2 (en) | 2010-07-30 | 2014-07-15 | Kla-Tencor Corp. | Dynamic care areas |
US20120031975A1 (en) * | 2010-08-04 | 2012-02-09 | The Code Corporation | Illumination blocks for a graphical code reader |
US9170211B2 (en) | 2011-03-25 | 2015-10-27 | Kla-Tencor Corp. | Design-based inspection using repeating structures |
US9087367B2 (en) | 2011-09-13 | 2015-07-21 | Kla-Tencor Corp. | Determining design coordinates for wafer defects |
US8831334B2 (en) | 2012-01-20 | 2014-09-09 | Kla-Tencor Corp. | Segmentation for wafer inspection |
US8826200B2 (en) | 2012-05-25 | 2014-09-02 | Kla-Tencor Corp. | Alteration for wafer inspection |
CN103674974A (en) * | 2012-09-06 | 2014-03-26 | 株式会社岛津制作所 | Inspection apparatus |
US9189844B2 (en) | 2012-10-15 | 2015-11-17 | Kla-Tencor Corp. | Detecting defects on a wafer using defect-specific information |
US9053527B2 (en) | 2013-01-02 | 2015-06-09 | Kla-Tencor Corp. | Detecting defects on a wafer |
US9134254B2 (en) | 2013-01-07 | 2015-09-15 | Kla-Tencor Corp. | Determining a position of inspection system output in design data space |
US9311698B2 (en) | 2013-01-09 | 2016-04-12 | Kla-Tencor Corp. | Detecting defects on a wafer using template image matching |
US9092846B2 (en) | 2013-02-01 | 2015-07-28 | Kla-Tencor Corp. | Detecting defects on a wafer using defect-specific and multi-channel information |
US9865512B2 (en) | 2013-04-08 | 2018-01-09 | Kla-Tencor Corp. | Dynamic design attributes for wafer inspection |
US9310320B2 (en) | 2013-04-15 | 2016-04-12 | Kla-Tencor Corp. | Based sampling and binning for yield critical defects |
JP2016035405A (en) * | 2014-08-01 | 2016-03-17 | リコーエレメックス株式会社 | Image inspection device, image inspection system, and image inspection method |
US10401411B2 (en) | 2014-09-29 | 2019-09-03 | Ecoatm, Llc | Maintaining sets of cable components used for wired analysis, charging, or other interaction with portable electronic devices |
US10438174B2 (en) | 2014-10-02 | 2019-10-08 | Ecoatm, Llc | Application for device evaluation and other processes associated with device recycling |
US11126973B2 (en) | 2014-10-02 | 2021-09-21 | Ecoatm, Llc | Wireless-enabled kiosk for recycling consumer devices |
US10496963B2 (en) | 2014-10-02 | 2019-12-03 | Ecoatm, Llc | Wireless-enabled kiosk for recycling consumer devices |
US11790327B2 (en) | 2014-10-02 | 2023-10-17 | Ecoatm, Llc | Application for device evaluation and other processes associated with device recycling |
US9911102B2 (en) | 2014-10-02 | 2018-03-06 | ecoATM, Inc. | Application for device evaluation and other processes associated with device recycling |
US10475002B2 (en) | 2014-10-02 | 2019-11-12 | Ecoatm, Llc | Wireless-enabled kiosk for recycling consumer devices |
US11734654B2 (en) | 2014-10-02 | 2023-08-22 | Ecoatm, Llc | Wireless-enabled kiosk for recycling consumer devices |
US10445708B2 (en) | 2014-10-03 | 2019-10-15 | Ecoatm, Llc | System for electrically testing mobile devices at a consumer-operated kiosk, and associated devices and methods |
US11232412B2 (en) | 2014-10-03 | 2022-01-25 | Ecoatm, Llc | System for electrically testing mobile devices at a consumer-operated kiosk, and associated devices and methods |
US10417615B2 (en) | 2014-10-31 | 2019-09-17 | Ecoatm, Llc | Systems and methods for recycling consumer electronic devices |
US10572946B2 (en) | 2014-10-31 | 2020-02-25 | Ecoatm, Llc | Methods and systems for facilitating processes associated with insurance services and/or other services for electronic devices |
US11436570B2 (en) | 2014-10-31 | 2022-09-06 | Ecoatm, Llc | Systems and methods for recycling consumer electronic devices |
US10860990B2 (en) | 2014-11-06 | 2020-12-08 | Ecoatm, Llc | Methods and systems for evaluating and recycling electronic devices |
US11080672B2 (en) | 2014-12-12 | 2021-08-03 | Ecoatm, Llc | Systems and methods for recycling consumer electronic devices |
US11315093B2 (en) | 2014-12-12 | 2022-04-26 | Ecoatm, Llc | Systems and methods for recycling consumer electronic devices |
US20160267339A1 (en) * | 2015-03-11 | 2016-09-15 | Canon Kabushiki Kaisha | Image processing apparatus and method of generating face image |
US10558851B2 (en) * | 2015-03-11 | 2020-02-11 | Canon Kabushiki Kaisha | Image processing apparatus and method of generating face image |
JP2017032548A (en) * | 2015-06-26 | 2017-02-09 | コグネックス・コーポレイション | Using 3d vision for automated industrial inspection |
JP7037876B2 (en) | 2015-06-26 | 2022-03-17 | コグネックス・コーポレイション | Use of 3D vision in automated industrial inspection |
TWI626438B (en) * | 2015-09-30 | 2018-06-11 | 佳能股份有限公司 | Inspection apparatus and article manufacturing method |
JP2017122682A (en) * | 2016-01-08 | 2017-07-13 | 株式会社キーエンス | Visual inspection device, visual inspection method, and computer program executable by controller used in visual inspection device |
US10127647B2 (en) | 2016-04-15 | 2018-11-13 | Ecoatm, Llc | Methods and systems for detecting cracks in electronic devices |
US9885672B2 (en) | 2016-06-08 | 2018-02-06 | ecoATM, Inc. | Methods and systems for detecting screen covers on electronic devices |
US10909673B2 (en) | 2016-06-28 | 2021-02-02 | Ecoatm, Llc | Methods and systems for detecting cracks in illuminated electronic device screens |
US11803954B2 (en) | 2016-06-28 | 2023-10-31 | Ecoatm, Llc | Methods and systems for detecting cracks in illuminated electronic device screens |
US10269110B2 (en) | 2016-06-28 | 2019-04-23 | Ecoatm, Llc | Methods and systems for detecting cracks in illuminated electronic device screens |
EP3847446A4 (en) * | 2018-09-06 | 2022-06-01 | Orbotech Ltd. | Multimodality multiplexed illumination for optical inspection systems |
WO2020049551A1 (en) | 2018-09-06 | 2020-03-12 | Orbotech Ltd. | Multimodality multiplexed illumination for optical inspection systems |
US11482067B2 (en) | 2019-02-12 | 2022-10-25 | Ecoatm, Llc | Kiosk for evaluating and purchasing used electronic devices |
US11462868B2 (en) | 2019-02-12 | 2022-10-04 | Ecoatm, Llc | Connector carrier for electronic device kiosk |
US11843206B2 (en) | 2019-02-12 | 2023-12-12 | Ecoatm, Llc | Connector carrier for electronic device kiosk |
US11798250B2 (en) | 2019-02-18 | 2023-10-24 | Ecoatm, Llc | Neural network based physical condition evaluation of electronic devices, and associated systems and methods |
US11974046B2 (en) | 2019-08-27 | 2024-04-30 | Orbotech Ltd. | Multimodality multiplexed illumination for optical inspection systems |
US20220292665A1 (en) * | 2019-10-02 | 2022-09-15 | Konica Minolta, Inc. | Workpiece surface defect detection device and detection method, workpiece surface inspection system, and program |
US20210350495A1 (en) * | 2020-02-10 | 2021-11-11 | Cognex Corporation | Composite three-dimensional blob tool and method for operating the same |
US11966996B2 (en) * | 2020-02-10 | 2024-04-23 | Cognex Corporation | Composite three-dimensional blob tool and method for operating the same |
US11922467B2 (en) | 2020-08-17 | 2024-03-05 | ecoATM, Inc. | Evaluating an electronic device using optical character recognition |
US20230084949A1 (en) * | 2021-09-13 | 2023-03-16 | Quanta Computer Inc. | Systems and methods for monitoring antenna arrays |
US11742962B2 (en) * | 2021-09-13 | 2023-08-29 | Quanta Computer Inc. | Systems and methods for monitoring antenna arrays |
Also Published As
Publication number | Publication date |
---|---|
WO2002099403A1 (en) | 2002-12-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20020186878A1 (en) | System and method for multiple image analysis | |
US20220084183A1 (en) | Defect detection device, defect detection method, and program | |
US9552636B2 (en) | Detecting defects on a wafer using defect-specific and multi-channel information | |
US6888959B2 (en) | Method of inspecting a semiconductor device and an apparatus thereof | |
US7177458B1 (en) | Reduction of false alarms in PCB inspection | |
CA2638415C (en) | Patterned wafer defect inspection system and method | |
US7420542B2 (en) | Apparatus for capturing and analyzing light and method embodied therein | |
US20030117616A1 (en) | Wafer external inspection apparatus | |
US7024031B1 (en) | System and method for inspection using off-angle lighting | |
JPH06307833A (en) | Surface unevenness shape recognizing device | |
JP2002296192A (en) | Method for inspecting flaw using color illumination | |
JP2009264882A (en) | Visual inspection device | |
JP2002250700A (en) | Method and device for inspecting pattern | |
KR102015620B1 (en) | System and Method for detecting Metallic Particles | |
JPH11352073A (en) | Foreign matter inspection method and apparatus thereof | |
Katafuchi et al. | A method for inspecting industrial parts surfaces based on an optics model | |
JP2004212218A (en) | Sample inspection method and inspection apparatus | |
JPS61193007A (en) | Inspecting method for rod type projection body | |
JP2003057193A (en) | Foreign matter checking apparatus | |
US20240094145A1 (en) | Detection method and system for determining the location of a surface defect on a front or back surface of a transparent film | |
JPH0413953A (en) | Detect inspection preprocessor for electronic component molded form | |
JP2532513B2 (en) | Object presence inspection method | |
KR100564871B1 (en) | Inspecting method and apparatus for repeated micro-miniature patterns | |
Kim | 3-Dimensional Micro Solder Ball Inspection Using LED Reflection Image | |
CN113447485A (en) | Optical detection method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: ASTI HOLDINGS, INC., SINGAPORE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HOON, TAN SEOW;RAO, SREENIVAS;REEL/FRAME:012439/0766;SIGNING DATES FROM 20010829 TO 20010921 |
|
AS | Assignment |
Owner name: ASTI HOLDINGS LIMITED, SINGAPORE Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE RECEIVING PARTY ADDRESS PREVIOUSLY RECORDED ON REEL 012429 FRAME 0766;ASSIGNORS:HOON, TAN SEOW;SREENIVAS, RAO;REEL/FRAME:012968/0341;SIGNING DATES FROM 20010829 TO 20010921 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |