WO2013100223A1 - 기판 검사장치의 높이정보 생성 방법 - Google Patents
기판 검사장치의 높이정보 생성 방법 Download PDFInfo
- Publication number
- WO2013100223A1 WO2013100223A1 PCT/KR2011/010243 KR2011010243W WO2013100223A1 WO 2013100223 A1 WO2013100223 A1 WO 2013100223A1 KR 2011010243 W KR2011010243 W KR 2011010243W WO 2013100223 A1 WO2013100223 A1 WO 2013100223A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- region
- image
- area
- height information
- matching
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05K—PRINTED CIRCUITS; CASINGS OR CONSTRUCTIONAL DETAILS OF ELECTRIC APPARATUS; MANUFACTURE OF ASSEMBLAGES OF ELECTRICAL COMPONENTS
- H05K13/00—Apparatus or processes specially adapted for manufacturing or adjusting assemblages of electric components
- H05K13/08—Monitoring manufacture of assemblages
- H05K13/089—Calibration, teaching or correction of mechanical systems, e.g. of the mounting head
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/02—Measuring arrangements characterised by the use of optical techniques for measuring length, width or thickness
- G01B11/06—Measuring arrangements characterised by the use of optical techniques for measuring length, width or thickness for measuring thickness ; e.g. of sheet material
- G01B11/0608—Height gauges
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/95—Investigating the presence of flaws or contamination characterised by the material or shape of the object to be examined
- G01N21/956—Inspecting patterns on the surface of objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0004—Industrial image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05K—PRINTED CIRCUITS; CASINGS OR CONSTRUCTIONAL DETAILS OF ELECTRIC APPARATUS; MANUFACTURE OF ASSEMBLAGES OF ELECTRICAL COMPONENTS
- H05K13/00—Apparatus or processes specially adapted for manufacturing or adjusting assemblages of electric components
- H05K13/08—Monitoring manufacture of assemblages
- H05K13/081—Integration of optical monitoring devices in assembly lines; Processes using optical monitoring devices specially adapted for controlling devices or machines in assembly lines
-
- H—ELECTRICITY
- H05—ELECTRIC TECHNIQUES NOT OTHERWISE PROVIDED FOR
- H05K—PRINTED CIRCUITS; CASINGS OR CONSTRUCTIONAL DETAILS OF ELECTRIC APPARATUS; MANUFACTURE OF ASSEMBLAGES OF ELECTRICAL COMPONENTS
- H05K13/00—Apparatus or processes specially adapted for manufacturing or adjusting assemblages of electric components
- H05K13/08—Monitoring manufacture of assemblages
- H05K13/081—Integration of optical monitoring devices in assembly lines; Processes using optical monitoring devices specially adapted for controlling devices or machines in assembly lines
- H05K13/0815—Controlling of component placement on the substrate during or after manufacturing
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/95—Investigating the presence of flaws or contamination characterised by the material or shape of the object to be examined
- G01N21/956—Inspecting patterns on the surface of objects
- G01N2021/95638—Inspecting patterns on the surface of objects for PCB's
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
- G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
- G01N21/84—Systems specially adapted for particular applications
- G01N21/88—Investigating the presence of flaws or contamination
- G01N21/8806—Specially adapted optical and illumination features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30108—Industrial image inspection
- G06T2207/30141—Printed circuit board [PCB]
Definitions
- the present invention relates to a method of generating height information of a substrate inspection apparatus, and more particularly, to a method of generating height information of a substrate inspection apparatus for inspecting a substrate through two or more camera photographing.
- At least one printed circuit board is provided in an electronic device, and various circuit elements such as a circuit pattern, a connection pad part, and a driving chip electrically connected to the connection pad part are provided on the printed circuit board. Are mounted.
- a substrate inspection apparatus is used to confirm that the various circuit elements as described above are properly formed or disposed on the printed circuit board.
- Conventional substrate inspection apparatus is generally a structure for taking an image using one camera, but recently, a structure for taking an image using two cameras is also used to increase the resolution of the image and to improve the inspection speed.
- the problem to be solved by the present invention is to accurately match the plurality of photographed images when inspecting the substrate by using a plurality of cameras, a substrate inspection apparatus that can accurately measure the height while utilizing the advantage of using a plurality of cameras It is to provide a height information generation method of.
- a first region of the calibration specimen and a second region at least partially overlapping with the first region are photographed.
- the first image and the second image are matched based on the overlapping area of the first area and the second area.
- a relative positional relationship between the first image and the second image is obtained using the matching.
- a first grating image corresponding to the first area and a second grating image corresponding to the second area are photographed by taking a measurement object formed on the test substrate based on the positional relationship using grating pattern light. Create merge height information.
- the calibration specimen may be marked such that a plurality of dots are arranged.
- An identifier may be formed on the calibration specimen, and the identifier is included in the overlapping area.
- the matching of the first image and the second image based on the overlapping area of the first area and the second area may include a feature dot that is a reference of the matching among the plurality of dots ( selecting feature dots within the overlapping area and matching the feature dots shown in the first image with the feature dots shown in the second image.
- the plurality of dots marked on the calibration specimen may be arranged in a matrix, and the overlapping area may include a matching line including at least one row of the matrix.
- the first image and the second image may be matched based on the matching line.
- a first image corresponding to the first region and a second image corresponding to the second region may be photographed by photographing the first region and the second region at least partially overlapping the first region.
- the method may further include masking the remaining dots of the plurality of dots arranged in the matrix form so as not to photograph the dots except for the matching line.
- Matching the first image and the second image based on the overlapping area of the first area and the second area may be performed by using an arrangement form of dots shown in the first image in which the first area is photographed.
- the method may include generating a prediction image corresponding to the second region connected to the first image, and matching the prediction image and the second image based on the overlapping region.
- It may include obtaining s, m11, m12, m13, m21, m22, m23, m31, m32, m33 satisfying the.
- Matching the first image and the second image based on the overlapping area of the first area and the second area may match at least five points to obtain the equation.
- the grid pattern light is irradiated to the measurement object formed on the inspection substrate.
- a first grid image corresponding to the first area and a second grid image corresponding to the second area are photographed by photographing a first area of the measurement object and a second area at least partially overlapping the first area.
- the obtained first grid image and the second grid image by using a bucket algorithm (bucket algorithm) the first height information corresponding to the first grid image and the second height corresponding to the second grid image Obtain information.
- the first height information and the second height information are merged using a relative positional relationship between the first area and the second area, which are obtained by using the overlapped area of the first area and the second area. To generate merge height information.
- the first height information may include a first measurement height H1
- the second height information may include a second measurement height H2, and an overlapping area between the first area and the second area.
- the grid pattern light is first irradiated to the measurement object formed on the inspection substrate. Subsequently, a first grid image corresponding to the first area and a second grid image corresponding to the second area are photographed by photographing a first area of the measurement object and a second area at least partially overlapping the first area. Acquire. Next, the first grating image and the second grating image are obtained by using a relative positional relationship between the first area and the second area, which are obtained by using an overlapping area of the first area and the second area. Create a merge grid image by merging. Subsequently, merge height information corresponding to the obtained merged grid image is obtained using a bucket algorithm.
- the first grid image and the second grid image are merged using a relative positional relationship between the first region and the second region, which are obtained by using the overlapping region of the first region and the second region.
- Generating a successful candidate image, matching the first region and the second region by using the positional relationship, satisfies Equation H1 H2 + b in the overlapping region of the first region and the second region Acquiring b and generating the merged grid image by using the acquired b.
- the height information generating method of the substrate inspection apparatus may further include calculating a confidence index for the first grid image and the second grid image, wherein the confidence index is a signal pattern in which the grid pattern light is received. It may include at least one or more of the signal strength, modulation, visibility and signal-to-noise ratio of the, the confidence index of the first region and the second region is excellent One region may be corrected based on one region.
- a relative positional relationship between images obtained from a plurality of cameras is obtained and the images are obtained using the acquired positional relationship. Since matching, accurate image matching is possible.
- FIG. 1 is a conceptual diagram illustrating a substrate inspection apparatus according to an embodiment of the present invention.
- FIG. 2 is a conceptual diagram illustrating an image photographing unit of a substrate inspection apparatus according to an exemplary embodiment of the present invention.
- FIG. 3 is a flowchart illustrating a height information generating method of a substrate inspection apparatus according to an exemplary embodiment of the present invention.
- FIG. 4 is a plan view illustrating an example of a calibration specimen that is employed in the substrate inspection apparatus of FIG. 2.
- FIG. 5 is a plan view illustrating an embodiment of a method of matching a first image and a second image in the height information generating method of the substrate inspecting apparatus of FIG. 3.
- FIG. 6 is a plan view illustrating another embodiment of a method of matching a first image and a second image in the height information generating method of the substrate inspection apparatus of FIG. 3.
- FIG. 7 is a plan view illustrating another embodiment of a method of matching a first image and a second image in the height information generation method of the substrate inspection apparatus of FIG. 3.
- FIG. 8 is a flowchart illustrating a height information generating method of a substrate inspection apparatus according to another exemplary embodiment of the present invention.
- FIG. 9 is a flowchart illustrating a height information generating method of a substrate inspection apparatus according to still another exemplary embodiment of the present invention.
- Image capturing unit 110a the first camera
- AR1 first region
- AR2 second region
- first and second may be used to describe various components, but the components should not be limited by the terms. The terms are used only for the purpose of distinguishing one component from another.
- the first component may be referred to as the second component, and similarly, the second component may also be referred to as the first component.
- FIG. 1 is a conceptual diagram illustrating a substrate inspection apparatus according to an embodiment of the present invention.
- a substrate inspection apparatus may include a first lighting unit including a measurement stage unit 100, an image photographing unit 200, and first and second lighting devices 300 and 400;
- the second lighting unit 450 may include an image acquirer 500, a module controller 600, and a central controller 700.
- the measurement stage unit 100 may include a stage 110 for supporting the measurement object 10 and a stage transfer unit 120 for transferring the stage 110.
- the measurement object 10 is moved by the stage 110 with respect to the image capturing unit 200 and the first and second lighting devices 300 and 400.
- the measurement position at) can be changed.
- the image capturing unit 200 is disposed above the stage 110 and receives the light reflected from the measurement object 10 to measure an image of the measurement object 10. That is, the image capturing unit 200 receives the light emitted from the first and second lighting devices 300 and 400 and reflected from the measurement object 10 to capture a planar image of the measurement object 10. do.
- the image capturing unit 200 may include a camera 210, an imaging lens 220, a filter 230, and a circular lamp 240.
- the camera 210 receives the light reflected from the measurement object 10 to take a planar image of the measurement object 10.
- the imaging lens 220 is disposed under the camera 210 to form light reflected from the measurement object 10 in the camera 210.
- the filter 230 is disposed below the imaging lens 220 to filter the light reflected from the measurement object 10 to provide the imaging lens 220, and for example, a frequency filter, a color filter, and light. It may be made of any one of the intensity control filter.
- the circular lamp 240 may be disposed under the filter 230 to provide light to the measurement object 10 to capture a specific image such as a two-dimensional shape of the measurement object 10.
- the first lighting device 300 may be disposed to be inclined with respect to the stage 110 supporting the measurement object 10 on the right side of the image photographing unit 200, for example.
- the first lighting device 300 may include a first lighting unit 310, a first grating unit 320, a first grating transfer unit 330, and a first condensing lens 340.
- the first lighting unit 310 is composed of an illumination source and at least one lens to generate light
- the first grating unit 320 is disposed below the first lighting unit 310 to the first illumination
- the light generated in the unit 310 is changed into the first lattice pattern light having the lattice pattern.
- the first grating transfer unit 330 is connected to the first grating unit 320 to transfer the first grating unit 320, for example, one of the PZT (Piezoelectric) transfer unit or fine linear transfer unit It can be adopted.
- the first condenser lens 340 is disposed under the first grating unit 320 to condense the first grating pattern light emitted from the first grating unit 320 to the measurement object 10.
- the second lighting device 400 may be disposed to be inclined with respect to the stage 110 supporting the measurement object 10 on the left side of the image capturing unit 200.
- the second lighting device 400 may include a second lighting unit 410, a second grating unit 420, a second grating transfer unit 430, and a second condensing lens 440. Since the second lighting device 400 is substantially the same as the first lighting device 300 described above, detailed descriptions thereof will be omitted.
- the first lighting device 300 irradiates the N first grating pattern light to the measurement object 10 while the first grating transfer unit 330 moves the first grating unit 320 sequentially N times.
- the image capturing unit 200 may photograph the N first pattern images by sequentially applying the N first lattice pattern lights reflected from the measurement object 10.
- the second grid transfer unit 430 moves the second grid unit 420 sequentially N times, and the N second grid pattern lights are directed to the measurement object 10.
- the image capturing unit 200 may photograph the N second pattern images by sequentially applying the N second grating pattern lights reflected from the measurement object 10.
- N is a natural number, for example, may be 3 or 4.
- the number of the lighting devices may be three or more. . That is, the grid pattern light irradiated to the measurement object 10 may be irradiated from various directions, and various kinds of pattern images may be photographed. For example, when three lighting devices are arranged in an equilateral triangle shape around the image capturing unit 200, three grid pattern lights may be applied to the measurement object 10 in different directions, and four When the lighting devices are arranged in a square shape around the image capturing unit 200, four grid pattern lights may be applied to the measurement object 10 in different directions.
- the first lighting unit may include eight lighting devices. In this case, the first lighting unit may photograph an image by radiating grid pattern light in eight directions.
- the second lighting unit 450 irradiates the measurement object 10 with light for obtaining a two-dimensional image of the measurement object 10.
- the second lighting unit 450 may include a red light 452, a green light 454, and a blue light 456.
- the red light 452, the green light 454, and the blue light 456 are disposed in a circular shape on the upper part of the measurement object 10, and the red light, green light, Blue light may be irradiated and may be formed to have different heights as shown in FIG. 1.
- the image acquisition unit 500 is electrically connected to the camera 210 of the image capturing unit 200 to obtain and store pattern images by the first lighting unit from the camera 210.
- the image acquisition unit 500 acquires and stores two-dimensional images by the second lighting unit 450 from the camera 210.
- the image acquisition unit 500 includes an image system that receives and stores the N first pattern images and the N second pattern images photographed by the camera 210.
- the module controller 600 is electrically connected to and controlled by the measurement stage unit 100, the image capturing unit 200, the first lighting device 300, and the second lighting device 400.
- the module controller 600 includes, for example, a lighting controller, a grid controller, and a stage controller.
- the lighting controller generates light by controlling the first and second lighting units 310 and 410, respectively, and the grid controller controls the first and second grid transfer units 330 and 430, respectively.
- the second grid units 320 and 420 are moved.
- the stage controller may control the stage transfer unit 120 to move the stage 110 up, down, left, and right.
- the central control unit 700 is electrically connected to the image acquisition unit 500 and the module control unit 600 to control each. Specifically, the central control unit 700 receives the N first pattern images and the N second pattern images from the image system of the image acquisition unit 500, processes them, and processes the three-dimensional image of the object to be measured. The shape can be measured. In addition, the central controller 700 may control the lighting controller, the grid controller, and the stage controller of the module controller 600, respectively. As such, the central control unit may include an image processing board, a control board, and an interface board.
- the image capturing unit 200 of the substrate inspection apparatus as described above may include at least two cameras.
- FIG. 2 is a conceptual diagram illustrating an image photographing unit of a substrate inspection apparatus according to an exemplary embodiment of the present invention.
- the image capturing unit 200a of the apparatus for inspecting a substrate may include a first camera 110a, a second camera 120a, and a beam splitter to inspect the measurement object 10. (beam splitter) 130a.
- the image capturing unit 200a includes at least two cameras to obtain an image by dividing an image of the measurement target 10.
- the first camera 110a acquires at least a portion of the image of the measurement target 10
- the second camera 120a acquires at least a portion of the image of the measurement target 10.
- the beam splitter 130a partially transmits and partially reflects the light reflected from the measurement object 10 and transmits the light to the first and second cameras 110a and 110b.
- the two images may be combined to obtain an image of one measurement object 10. Can be.
- FIG. 3 is a flowchart illustrating a method of generating height information of a substrate inspection apparatus according to an exemplary embodiment of the present invention
- FIG. 4 is a plan view illustrating an example of a calibration specimen employed in the substrate inspection apparatus of FIG. 1
- FIG. 2 is a plan view illustrating an embodiment of a method of matching a first image and a second image in the height information generating method of the substrate inspecting apparatus of FIG. 2.
- the first region AR1 and the first region of the calibration specimen 50 are first.
- the first image IM1 corresponding to the first area AR1 and the second image IM2 corresponding to the second area AR2 are photographed by photographing the second area AR2 at least partially overlapping the AR1.
- the calibration specimen 50 is an object for performing a correction operation for setting a relative positional relationship between the first image IM1 and the second image IM2, and the positional relationship before the measurement object is measured. Are employed to obtain.
- the calibration specimen 50 may be marked such that a plurality of dots 52 are arranged as shown in FIG. 5.
- the plurality of dots 52 marked on the calibration specimen 50 may be arranged in a matrix form.
- An identifier may be formed in the calibration specimen 50, and the identifier ID is included in the overlapping area OA.
- the delimiter ID may be formed in various ways to prevent a matching error when matching the first image IM1 and the second image IM2 based on the overlap region OA. Accordingly, the delimiter ID may be formed in a manner of characterizing at least one of the dots 52 or may be a mark displayed between the dots 52. For example, the separator ID may be formed by increasing the size of one dot in the overlapping area OA, as shown in FIG. 5. Alternatively, the identifier ID may be a cross mark formed between the dots 52.
- the first image IM1 and the second image IM2 are matched based on the overlapping area OA of the first area AR1 and the second area AR2 (S120). .
- the first area AR1 and the second area AR2 correspond to two divided portions of the upper surface of the calibration specimen 50, and the overlapping area OA is the first area AR1 and the second area AR2. Since the area is common to the second area AR2, the first image IM1 and the second area AR2, which are captured images of the first area AR1, are based on the overlap area OA. When the second image IM2, which is a captured image, is matched, a merged image of the upper surface of the calibration specimen 50 including the first and second regions AR1 and AR2 may be obtained.
- the separator ID when the separator ID is formed in the overlapping area OA, it may be more advantageous to match the first image IM1 and the second image IM2.
- a feature dot FD which is a reference of the matching, is selected within the overlap area OA.
- the feature dots FD may correspond to a part of a row formed on the calibration specimen 50 including a dot located at the rightmost side of the overlap area OA illustrated in FIG. 5. May be dots.
- the feature dots FD shown in the first image IM1 and the feature dots FD shown in the second image IM2 are matched.
- the separator ID when the separator ID is formed in the overlapping area OA, it may be more advantageous to match the first image IM1 and the second image IM2. have.
- FIG. 6 is a plan view illustrating another embodiment of a method of matching a first image and a second image in the height information generating method of the substrate inspection apparatus of FIG. 3.
- the overlapping area OA may include a matching line ML including at least one line of the matrix in which the plurality of dots 52 are arranged.
- the first image IM1 and the second image IM2 may be matched based on the matching line ML.
- the matching line ML is excluded from the plurality of dots 52 arranged in the matrix form. The remaining dots may be hidden from being photographed, thereby preventing a matching error.
- the obscured portions are shown as diagonal lines.
- FIG. 7 is a plan view illustrating another embodiment of a method of matching a first image and a second image in the height information generation method of the substrate inspection apparatus of FIG. 3.
- first image IM1 and the second image IM2 are matched based on the overlapping area OA of the first area AR1 and the second area AR2 ( S120) may be performed as follows.
- the prediction image PI is a virtual image that is naturally derived from the first image IM1 purely. Therefore, the prediction image PI corresponds to an image in which the first image IM1 is extended.
- the prediction image PI and the second image IM2 are matched based on the overlap region OA.
- the above-described delimiter (ID), the matching line, etc. may be applied as it is.
- Arrows shown in FIG. 7 exemplarily show matching relationships between dots.
- the positional relationship may be obtained by using the following equation.
- the positional relationship can be defined.
- s corresponds to a scale factor.
- Equation 1 Since point 1 is a point on a two-dimensional plane, with reference to Equation 1, one equation for u1, u2, and v2 and another equation for v1, u2, and v2 can be obtained. In addition, s may be represented for u2 and v2.
- a first lattice image corresponding to the first area AR1 and a second lattice image corresponding to the second area AR2 obtained by photographing the measurement object formed on the test substrate based on the positional relationship using the grid pattern light. Merging the two grid images to generate the merge height information (S140).
- merging height information may be generated using the merged grid image.
- FIG. 8 is a flowchart illustrating a height information generating method of a substrate inspection apparatus according to another exemplary embodiment of the present invention.
- a grid pattern light is irradiated to a measurement object formed on an inspection substrate (S210).
- a first grid image corresponding to the first area and a second grid image corresponding to the second area are photographed by photographing a first area of the measurement object and a second area at least partially overlapping the first area.
- Acquire S220.
- the obtained first grid image and the second grid image by using a bucket algorithm (bucket algorithm) the first height information corresponding to the first grid image and the second height corresponding to the second grid image
- Obtain information S230.
- the first height information and the second height information are merged using a relative positional relationship between the first area and the second area, which are obtained by using the overlapped area of the first area and the second area.
- merge height information S240.
- the first height information may include a first measurement height H1
- the second height information may include a second measurement height H2.
- the first region and the second region are matched using the positional relationship, and then the following equation is obtained in the overlapping region of the first region and the second region.
- a is a scale factor and b is a height difference.
- H1 means first height information
- H2 means second height information.
- the a and b may be obtained by substituting a plurality of H1 and H2 for the overlapping area and utilizing a least square method.
- the merge height information may be generated using the obtained a and b.
- FIG. 9 is a flowchart illustrating a height information generating method of a substrate inspection apparatus according to still another exemplary embodiment of the present invention.
- a grid pattern light is irradiated to a measurement object formed on an inspection substrate (S310).
- a first grid image corresponding to the first area and a second grid image corresponding to the second area are photographed by photographing a first area of the measurement object and a second area at least partially overlapping the first area.
- Acquire (S320) the first grating image and the second grating image are obtained by using a relative positional relationship between the first area and the second area, which are obtained by using an overlapping area of the first area and the second area.
- Merging generates a merge grid image (S330).
- merge height information corresponding to the obtained merged grid image is obtained using a bucket algorithm (S340).
- This method is generally similar to the process of FIG. 8 described above, but obtains a merged grid image without first obtaining respective heights from the first grid image and the second grid image, and then merges directly from the obtained merged grid image. There is a difference in generating height information.
- the first grid image and the second grid image are merged using a relative positional relationship between the first region and the second region, which are obtained by using the overlapping region of the first region and the second region.
- it may be performed in the following manner.
- the first region and the second region are matched using the positional relationship, and then the following equation is obtained in the overlapping region of the first region and the second region.
- H1 means first height information
- H2 means second height information
- b may be obtained by substituting a plurality of H1 and H2 for the overlapping area by an average value, a representative value, and a mode.
- the merged grid image may be generated using the obtained b.
- a confidence index for the first grid image and the second grid image may be calculated and used.
- the confidence index may include at least one of signal intensity, modulation, visibility, and signal-to-noise ratio of the signal pattern in which the grid pattern light is received.
- the other one of the first region and the second region may be corrected based on one region having the excellent confidence index.
Landscapes
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Microelectronics & Electronic Packaging (AREA)
- Operations Research (AREA)
- Manufacturing & Machinery (AREA)
- Quality & Reliability (AREA)
- Life Sciences & Earth Sciences (AREA)
- Analytical Chemistry (AREA)
- Chemical & Material Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Pathology (AREA)
- Immunology (AREA)
- Health & Medical Sciences (AREA)
- Biochemistry (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Investigating Materials By The Use Of Optical Means Adapted For Particular Applications (AREA)
- Image Processing (AREA)
Abstract
Description
Claims (14)
- 보정용 시편의 제1 영역 및 상기 제1 영역과 적어도 일부가 중첩되는 제2 영역을 촬영하여 상기 제1 영역에 대응하는 제1 이미지 및 상기 제2 영역에 대응하는 제2 이미지를 획득하는 단계;상기 제1 영역 및 상기 제2 영역의 중첩영역을 기초로 상기 제1 이미지 및 상기 제2 이미지를 매칭(matching)시키는 단계;상기 매칭을 이용하여 상기 제1 이미지와 상기 제2 이미지 사이의 상대적인 위치관계를 획득하는 단계; 및상기 위치관계를 기초로 검사기판 상에 형성된 측정대상물을 격자패턴광을 이용하여 촬영한 상기 제1 영역에 대응하는 제1 격자이미지 및 상기 제2 영역에 대응하는 제2 격자이미지를 병합하여 병합높이정보를 생성하는 단계를 포함하는 기판 검사장치의 높이정보 생성 방법.
- 제1항에 있어서,상기 보정용 시편에는 다수의 도트(dot)들이 배열되도록 마킹(marking)되어 있는 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
- 제2항에 있어서,상기 보정용 시편에는 구분자(identifier)가 형성되고,상기 구분자는 상기 중첩영역 내에 포함된 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
- 제2항에 있어서,상기 제1 영역 및 상기 제2 영역의 중첩영역을 기초로 상기 제1 이미지 및 상기 제2 이미지를 매칭시키는 단계는,상기 다수의 도트들 중에서 상기 매칭의 기준이 되는 특징도트(feature dot)들을 상기 중첩영역 내에서 선정하는 단계; 및상기 제1 이미지에 나타난 상기 특징도트들과 상기 제2 이미지에 나타난 상기 특징도트들을 매칭시키는 단계를 포함하는 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
- 제2항에 있어서,상기 보정용 시편에 마킹된 다수의 도트들은 매트릭스(matrix) 형태로 배열되고,상기 중첩영역은 상기 매트릭스의 적어도 하나 이상의 행(row)으로 이루어진 매칭 라인을 포함하며,상기 제1 이미지 및 상기 제2 이미지는 상기 매칭 라인을 기준으로 매칭되는 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
- 제5항에 있어서,상기 보정용 시편의 제1 영역 및 상기 제1 영역과 적어도 일부가 중첩되는 제2 영역을 촬영하여 상기 제1 영역에 대응하는 제1 이미지 및 상기 제2 영역에 대응하는 제2 이미지를 획득하는 단계 이전에,상기 매트릭스 형태로 배열된 다수의 도트들 중 상기 매칭 라인을 제외한 나머지 도트들을 촬영되지 않도록 가리는 단계를 더 포함하는 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
- 제2항에 있어서,상기 제1 영역 및 상기 제2 영역의 중첩영역을 기초로 상기 제1 이미지 및 상기 제2 이미지를 매칭시키는 단계는,상기 제1 영역이 촬영된 상기 제1 이미지에 나타난 도트들의 배열 형태를 이용하여 상기 제1 이미지에 연결되는 상기 제2 영역에 대응하는 예측 이미지를 생성하는 단계; 및상기 예측 이미지 및 상기 제2 이미지를 상기 중첩영역을 기초로 매칭시키는 단계를 포함하는 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
- 제8항에 있어서,상기 제1 영역 및 상기 제2 영역의 중첩영역을 기초로 상기 제1 이미지 및 상기 제2 이미지를 매칭시키는 단계는,상기 수학식을 획득하도록 적어도 5지점 이상을 매칭시키는 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
- 검사기판 상에 형성된 측정대상물에 격자패턴광을 조사하는 단계;상기 측정대상물의 제1 영역 및 상기 제1 영역과 적어도 일부가 중첩되는 제2 영역을 촬영하여 상기 제1 영역에 대응하는 제1 격자이미지 및 상기 제2 영역에 대응하는 제2 격자이미지를 획득하는 단계;획득된 상기 제1 격자이미지 및 상기 제2 격자이미지를 버킷 알고리즘(bucket algorithm)을 이용하여 상기 제1 격자이미지에 대응하는 제1 높이정보 및 상기 제2 격자이미지에 대응하는 제2 높이정보를 획득하는 단계; 및상기 제1 영역 및 상기 제2 영역의 중첩영역을 이용하여 이미 획득한 상기 제1 영역 및 상기 제2 영역 사이의 상대적인 위치관계를 이용하여 상기 제1 높이정보 및 상기 제2 높이정보를 병합하여 병합높이정보를 생성하는 단계를 포함하는 기판 검사장치의 높이정보 생성 방법.
- 제10항에 있어서,상기 제1 높이정보는 제1 측정높이(H1)를 포함하고, 상기 제2 높이정보는 제2 측정높이(H2)를 포함하며,상기 제1 영역 및 상기 제2 영역의 중첩영역을 이용하여 이미 획득한 상기 제1 영역 및 상기 제2 영역 사이의 상대적인 위치관계를 이용하여 상기 제1 높이정보 및 상기 제2 높이정보를 병합하여 병합높이정보를 생성하는 단계는,상기 위치관계를 이용하여 상기 제1 영역 및 상기 제2 영역을 매칭시키는 단계;상기 제1 영역 및 상기 제2 영역의 중첩영역에서 수학식 H1=aH2+b를 만족하는 a 및 b를 획득하는 단계; 및상기 획득된 a 및 b를 이용하여 상기 병합높이정보를 생성하는 단계를 포함하는 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
- 검사기판 상에 형성된 측정대상물에 격자패턴광을 조사하는 단계;상기 측정대상물의 제1 영역 및 상기 제1 영역과 적어도 일부가 중첩되는 제2 영역을 촬영하여 상기 제1 영역에 대응하는 제1 격자이미지 및 상기 제2 영역에 대응하는 제2 격자이미지를 획득하는 단계;상기 제1 영역 및 상기 제2 영역의 중첩영역을 이용하여 이미 획득한 상기 제1 영역 및 상기 제2 영역 사이의 상대적인 위치관계를 이용하여 상기 제1 격자이미지 및 상기 제2 격자이미지를 병합하여 병합격자이미지를 생성하는 단계; 및버킷 알고리즘을 이용하여 상기 획득된 병합격자이미지에 대응하는 병합높이정보를 획득하는 단계를 포함하는 기판 검사장치의 높이정보 생성 방법.
- 제12항에 있어서,상기 제1 영역 및 상기 제2 영역의 중첩영역을 이용하여 이미 획득한 상기 제1 영역 및 상기 제2 영역 사이의 상대적인 위치관계를 이용하여 상기 제1 격자이미지 및 상기 제2 격자이미지를 병합하여 병합격자이미지를 생성하는 단계는,상기 위치관계를 이용하여 상기 제1 영역 및 상기 제2 영역을 매칭시키는 단계;상기 제1 영역 및 상기 제2 영역의 중첩영역에서 수학식 H1=H2+b를 만족하는 b를 획득하는 단계; 및상기 획득된 b를 이용하여 상기 병합격자이미지를 생성하는 단계를 포함하는 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
- 제10항 및 제12항 중 어느 한 항에 있어서,상기 제1 격자이미지 및 상기 제2 격자이미지에 대한 신뢰지수를 산출하는 단계를 더 포함하고,상기 신뢰지수는 상기 격자패턴광이 수신된 신호패턴의 신호강도, 모듈레이션(modulation), 비저빌리티(visibility) 및 신호대잡음비(signal-to-noise ratio) 중 적어도 하나 이상을 포함하며,상기 제1 영역 및 상기 제2 영역 중 상기 신뢰지수가 우수한 어느 한 영역을 기준으로 나머지 한 영역이 보정되는 것을 특징으로 하는 기판 검사장치의 높이정보 생성 방법.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/347,137 US9115984B2 (en) | 2011-12-27 | 2011-12-28 | Method of generating height information in circuit board inspection apparatus |
JP2014534455A JP2014534420A (ja) | 2011-12-27 | 2011-12-28 | 基板検査装置の高さ情報生成方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020110143531A KR101215083B1 (ko) | 2011-12-27 | 2011-12-27 | 기판 검사장치의 높이정보 생성 방법 |
KR10-2011-0143531 | 2011-12-27 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2013100223A1 true WO2013100223A1 (ko) | 2013-07-04 |
Family
ID=47908198
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2011/010243 WO2013100223A1 (ko) | 2011-12-27 | 2011-12-28 | 기판 검사장치의 높이정보 생성 방법 |
Country Status (4)
Country | Link |
---|---|
US (1) | US9115984B2 (ko) |
JP (1) | JP2014534420A (ko) |
KR (1) | KR101215083B1 (ko) |
WO (1) | WO2013100223A1 (ko) |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102473547B1 (ko) * | 2018-02-26 | 2022-12-05 | 주식회사 고영테크놀러지 | 부품의 실장 상태를 검사하기 위한 방법, 인쇄 회로 기판 검사 장치 및 컴퓨터 판독 가능한 기록매체 |
JP7000380B2 (ja) * | 2019-05-29 | 2022-01-19 | Ckd株式会社 | 三次元計測装置及び三次元計測方法 |
CN113884509A (zh) * | 2020-06-16 | 2022-01-04 | 深圳中科飞测科技股份有限公司 | 检测装置及检测方法 |
CN112985333A (zh) * | 2021-02-19 | 2021-06-18 | 上海闻泰信息技术有限公司 | 电路板检查方法和装置 |
CN117091516B (zh) * | 2022-05-12 | 2024-05-28 | 广州镭晨智能装备科技有限公司 | 电路板保护层的厚度检测方法、***及存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004226128A (ja) * | 2003-01-20 | 2004-08-12 | Saki Corp:Kk | 外観検査装置および外観検査方法 |
JP2010169433A (ja) * | 2009-01-20 | 2010-08-05 | Ckd Corp | 三次元計測装置 |
KR101056995B1 (ko) * | 2009-07-03 | 2011-08-16 | 주식회사 고영테크놀러지 | 3차원 형상 검사방법 |
KR101059697B1 (ko) * | 2009-05-13 | 2011-08-29 | 주식회사 고영테크놀러지 | 인쇄회로기판상의 측정대상물의 측정방법 |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5416562A (en) * | 1992-03-06 | 1995-05-16 | Nikon Corporation | Method of detecting a position and apparatus therefor |
US20070131877A9 (en) * | 1999-11-29 | 2007-06-14 | Takashi Hiroi | Pattern inspection method and system therefor |
JP2002181729A (ja) * | 2000-12-12 | 2002-06-26 | Saki Corp:Kk | 外観検査装置および外観検査方法 |
CN100523795C (zh) * | 2001-11-30 | 2009-08-05 | 国际商业机器公司 | 图形轮廓的检查装置和检查方法、曝光装置 |
US7221788B2 (en) * | 2003-07-01 | 2007-05-22 | Infineon Technologies Ag | Method of inspecting a mask or reticle for detecting a defect, and mask or reticle inspection system |
US7545512B2 (en) * | 2006-01-26 | 2009-06-09 | Koh Young Technology Inc. | Method for automated measurement of three-dimensional shape of circuit boards |
JP2008014940A (ja) | 2006-06-08 | 2008-01-24 | Fast:Kk | 平面状被撮像物のカメラ計測のためのカメラキャリブレーション方法、および応用計測装置 |
JP2008191122A (ja) * | 2007-02-08 | 2008-08-21 | Toshiba Corp | 表面形状測定装置及び方法 |
US8379229B2 (en) * | 2008-02-14 | 2013-02-19 | Seiko Epson Corporation | Simulation of a printed dot-pattern bitmap |
TWI408486B (zh) * | 2008-12-30 | 2013-09-11 | Ind Tech Res Inst | 具動態校正的攝影機及其動態校正方法 |
DE102010029319B4 (de) * | 2009-05-27 | 2015-07-02 | Koh Young Technology Inc. | Vorrichtung zur Messung einer dreidimensionalen Form und Verfahren dazu |
US8331726B2 (en) * | 2009-06-29 | 2012-12-11 | International Business Machines Corporation | Creating emission images of integrated circuits |
DE102010030859B4 (de) * | 2009-07-03 | 2019-01-24 | Koh Young Technology Inc. | Verfahren zum Untersuchen eines auf einem Substrat montierten Messobjektes |
KR101207198B1 (ko) * | 2010-01-18 | 2012-12-03 | 주식회사 고영테크놀러지 | 기판 검사장치 |
-
2011
- 2011-12-27 KR KR1020110143531A patent/KR101215083B1/ko active IP Right Grant
- 2011-12-28 JP JP2014534455A patent/JP2014534420A/ja active Pending
- 2011-12-28 WO PCT/KR2011/010243 patent/WO2013100223A1/ko active Application Filing
- 2011-12-28 US US14/347,137 patent/US9115984B2/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004226128A (ja) * | 2003-01-20 | 2004-08-12 | Saki Corp:Kk | 外観検査装置および外観検査方法 |
JP2010169433A (ja) * | 2009-01-20 | 2010-08-05 | Ckd Corp | 三次元計測装置 |
KR101059697B1 (ko) * | 2009-05-13 | 2011-08-29 | 주식회사 고영테크놀러지 | 인쇄회로기판상의 측정대상물의 측정방법 |
KR101056995B1 (ko) * | 2009-07-03 | 2011-08-16 | 주식회사 고영테크놀러지 | 3차원 형상 검사방법 |
Also Published As
Publication number | Publication date |
---|---|
US9115984B2 (en) | 2015-08-25 |
US20140219542A1 (en) | 2014-08-07 |
KR101215083B1 (ko) | 2012-12-24 |
JP2014534420A (ja) | 2014-12-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2016200096A1 (ko) | 3차원 형상 측정장치 | |
WO2009107981A2 (ko) | 3차원형상 측정장치 및 측정방법 | |
WO2009142390A2 (ko) | 표면형상 측정장치 | |
WO2014163375A1 (ko) | 기판의 이물질 검사방법 | |
WO2011087337A2 (ko) | 기판 검사장치 | |
WO2016163840A1 (ko) | 3차원 형상 측정장치 | |
WO2012050378A2 (ko) | 기판 검사방법 | |
WO2013176482A1 (ko) | 3차원 형상 측정장치의 높이 측정 방법 | |
WO2013100223A1 (ko) | 기판 검사장치의 높이정보 생성 방법 | |
WO2013048093A2 (ko) | 비접촉식 부품검사장치 및 부품검사방법 | |
WO2015080480A1 (ko) | 웨이퍼 영상 검사 장치 | |
US20090147245A1 (en) | System and method for measuring optical resolution of lens | |
WO2012091494A2 (ko) | 기판 검사방법 | |
WO2017014518A1 (ko) | 검사 시스템 및 검사 방법 | |
WO2012023816A2 (ko) | 멀티 형광영상 관측용 형광현미경, 이를 이용한 형광영상의 관찰방법 및 멀티 형광영상 관측 시스템 | |
KR101241175B1 (ko) | 실장기판 검사장치 및 검사방법 | |
WO2013009065A2 (ko) | 엘이디 부품의 3차원비전검사장치 및 비전검사방법 | |
WO2016099154A1 (ko) | 부품이 실장된 기판 검사방법 및 검사장치 | |
WO2013183471A1 (ja) | 外観検査装置及び外観検査方法 | |
WO2012134146A1 (ko) | 스테레오 비전과 격자 무늬를 이용한 비전검사장치 | |
KR20050051535A (ko) | 결함 검사 장치 | |
WO2015026210A1 (ko) | 솔더 조인트 검사 방법 | |
JP2006245891A (ja) | カメラモジュールの画像検査用チャート、この画像検査用チャートを用いたカメラモジュールの画像検査方法および画像検査装置 | |
WO2017146300A1 (ko) | 마킹 위치 보정장치 및 방법 | |
WO2013009151A2 (ko) | 검사방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11878539 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14347137 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2014534455 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 11878539 Country of ref document: EP Kind code of ref document: A1 |