WO2010032792A1 - 3次元計測装置およびその方法 - Google Patents
3次元計測装置およびその方法 Download PDFInfo
- Publication number
- WO2010032792A1 WO2010032792A1 PCT/JP2009/066272 JP2009066272W WO2010032792A1 WO 2010032792 A1 WO2010032792 A1 WO 2010032792A1 JP 2009066272 W JP2009066272 W JP 2009066272W WO 2010032792 A1 WO2010032792 A1 WO 2010032792A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- dimensional
- cameras
- normal
- images
- coordinate system
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
- G01B11/245—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures using a plurality of fixed, simultaneously operating transducers
Definitions
- the present invention relates to a technique for measuring a three-dimensional shape of a measurement object, in particular, a measurement object of a mirror surface.
- Three-dimensional measurement is a technology that measures distance by examining the correspondence between pixels from images taken by multiple cameras with different imaging angles, as shown in FIG. 12, and calculating parallax. .
- the luminance value is used as a feature when examining the corresponding pixel.
- the luminance value captured in the image does not represent the feature amount of the object surface itself, but is determined by the reflection of the surrounding object. Therefore, as shown in FIG. 13, when a specular object is photographed by two cameras 101 and 102, the position of the object surface where the light emitted from the light source L1 is reflected is different. When three-dimensional surveying is performed with these points as corresponding pixels, in practice, the location of the point L2 in the figure is measured, and an error occurs. This error increases as the camera imaging angle difference increases. In addition, errors due to differences in camera characteristics also occur.
- the conventional three-dimensional survey method is applied to a mirror surface object using the luminance value as the feature amount, the difference in the characteristics of the multiple cameras and the arrangement of the cameras affect the luminance value of the captured image. I will. When the surface to be measured is a mirror surface, this effect is increased.
- Patent Document 1 focuses on the information that is specific to the measurement target, that is, the normal, so that errors due to differences in camera placement and characteristics can be reduced, but the errors due to area division It occurs.
- the surface resolution is coarse due to area division, and it can only be measured as an angular three-dimensional shape.
- the convergence angle of the camera is reduced and the same coordinate system is regarded as having the same coordinate system among the plurality of cameras when the association is performed, the accuracy of the association is different according to the difference of the normal coordinate system when the convergence angle is increased. Will decrease.
- the present invention has been made in view of the above-described circumstances, and the object of the present invention is a technique capable of accurately measuring the three-dimensional shape of a specular object without being affected by differences in camera position and characteristics. To provide.
- a three-dimensional measuring apparatus is a three-dimensional measuring apparatus for measuring a three-dimensional shape of a measurement object which is a specular object, and is photographed by a plurality of cameras and the plurality of cameras.
- Feature acquiring means for acquiring physical features of the surface of the measurement object from each of the captured images, and corresponding pixels for searching for corresponding pixels among the images captured by the plurality of cameras using the physical features It is characterized by comprising search means and surveying means for performing three-dimensional survey based on the parallax between corresponding pixels.
- An error occurs when taking pixel correspondence using luminance information reflected on the surface of a specular object, because the information is not information of the surface of the specular object itself, but changes according to conditions such as ambient illumination. is there. Therefore, in the present invention, by acquiring the physical feature of the surface of the specular object and taking pixel correspondence using this feature, high-precision matching is possible without being affected by the position and orientation of the camera. Therefore, it is possible to measure the three-dimensional shape of the measurement object with high accuracy.
- the direction of the surface normal is a physical feature of the surface of the measurement object as described above.
- spectral characteristics or reflection characteristics of the surface of the measurement object may be used. All of these physical features are information specific to the measurement object, and thus are not influenced by the position or orientation of the camera.
- the corresponding pixel search means searches for corresponding pixels among the images by using the direction of the normal which is converted into the common coordinate system by the coordinate conversion means.
- the corresponding pixel search means in the present invention compare physical features in an area of a predetermined width including the pixel of interest to search for corresponding pixels among the images. By performing comparison including surrounding physical features, it is possible to perform more accurate matching.
- the present invention can be grasped as a three-dimensional measurement device having at least a part of the above-mentioned means. Furthermore, the present invention can also be regarded as a three-dimensional measurement method including at least a part of the above-described processing, or a program for realizing such a method. Each of the above means and processes can be combined with one another as much as possible to constitute the present invention.
- the present invention it is possible to measure the three-dimensional shape of a specular object with high accuracy without being affected by the difference in the position or the characteristic of the camera.
- FIG. 1 is a diagram showing an outline of a three-dimensional measurement apparatus.
- FIG. 2 is a diagram showing functional blocks of the three-dimensional measurement apparatus.
- FIG. 3 is a diagram for explaining the arrangement of the cameras.
- FIG. 4A is a view for explaining the azimuthal arrangement of the lighting device.
- FIG. 4B is a view for explaining the zenith angle arrangement of the lighting device.
- FIG. 5 is a diagram showing a functional block diagram of the surface shape calculation unit.
- FIG. 6 is a diagram for explaining a method of creating a normal-luminance table.
- FIG. 7 is a diagram for explaining a method of acquiring the direction of the normal from the photographed image.
- FIG. 8 is a diagram for explaining a transformation matrix for transforming between the world coordinate system and the coordinate system of each camera.
- FIG. 9 is a flowchart showing the flow of corresponding point search processing by the corresponding point calculation unit.
- FIG. 10A is a diagram for explaining a search window for the corresponding point search.
- FIG. 10B is a diagram for explaining calculation of similarity for searching for corresponding points.
- FIG. 11 is a diagram for explaining a lighting device in the second embodiment.
- FIG. 12 is a diagram showing the principle of three-dimensional surveying.
- FIG. 13 is a diagram for explaining the case where a three-dimensional survey is performed on a specular object.
- FIG. 1 is a view showing an outline of a three-dimensional measurement apparatus according to the present embodiment.
- FIG. 2 is a diagram showing functional blocks of the three-dimensional measurement apparatus according to the present embodiment.
- the measurement object 4 disposed on the stage 5 is photographed by two cameras 1 and 2.
- White light is emitted to the measurement object 4 from three illumination devices 3a to 3c from different directions, and the illumination devices 3a to 3c are sequentially illuminated, and the cameras 1 and 2 respectively receive three images.
- the photographed image is taken into the computer 6, subjected to image processing, and three-dimensional measurement is performed.
- the computer 6 functions as the surface shape calculation unit 7, the coordinate conversion unit 8, the corresponding point calculation unit 9, and the triangulation unit 10 as shown in FIG. Note that some or all of these functional units may be realized by dedicated hardware.
- FIG. 3 is a diagram for explaining the arrangement of the cameras. As shown in FIG. 3, the camera 1 shoots the measurement target 4 from the vertical direction, and the camera 2 shoots the measurement target 4 from the direction shifted 40 degrees from the vertical direction.
- FIG. 4 is a diagram for explaining the arrangement of the lighting devices 3a to 3c.
- FIG. 4A is a view as viewed from the vertical direction, showing an azimuthal arrangement of the lighting devices 3a to 3c.
- FIG. 4B is a view as seen from the lateral direction, showing a zenith angle arrangement of the lighting devices 3a to 3c.
- the three illumination devices 3a to 3c irradiate light to the object to be measured from directions with azimuth angles different from each other by 120 degrees, and from directions with a zenith angle of 40 degrees.
- the arrangement of the cameras 1 and 2 and the illumination devices 3a to 3c described here is only one specific example, and the arrangement does not necessarily have to be such.
- the azimuth angles of the lighting devices may not be uniform.
- the zenith angle of a camera and an illuminating device is made the same, you may differ.
- the surface shape calculation unit 7 is a functional unit that calculates the direction of the normal at each position of the measurement object from the three images captured by the cameras 1 and 2 respectively.
- a more detailed functional block diagram of the surface shape calculation unit 7 is shown in FIG. As shown in the figure, the surface shape calculation unit 7 has an image input unit 71, a normal-brightness table 72, and a normal calculation unit 73.
- the image input unit 71 is a functional unit that receives an input of an image captured by the cameras 1 and 2.
- the image input unit 71 converts analog data into digital data by a capture board or the like when receiving analog data from the cameras 1 and 2.
- the image input unit 71 may also receive an image of digital data via a USB terminal or an IEEE 1394 terminal.
- a configuration may be adopted in which an image is read from a portable storage medium via a LAN cable.
- the normal-brightness table 72 is a storage unit that stores the correspondence between the direction of the normal and the brightness value of the image captured by sequentially lighting the three lighting devices 3a to 3c.
- the normal-brightness table 72 is prepared for each camera, and in the present embodiment, two normal-brightness tables corresponding to the cameras 1 and 2 are used.
- a method of creating the normal-brightness table 72 will be described with reference to FIG.
- the illumination devices 3a to 3c are sequentially turned on to capture three images 10a to 10c for an object whose surface shape is known.
- it is preferable to target the sphere because the sphere has normals in all directions and the direction of the normal at each position can be easily calculated by calculation.
- the direction of the normal (zenith angle ⁇ and azimuth angle ⁇ ) and the luminance value (La, Lb, Lc) in each image are acquired for each position of the table creation images 10a to 10c, and these are corresponded. And memorize.
- the normal-brightness table 72 By associating the combination of the direction of the normal with the luminance value at every point in the photographed image, it is possible to create the normal-brightness table 72 storing the combination of the direction of the normal and the luminance for all the directions of the normal. .
- the normal line calculation unit 73 calculates the direction of the normal line at each position of the measurement object 4 from the three images 11a to 11c photographed by sequentially lighting the lighting devices 3a to 3c as shown in FIG. Do. More specifically, the normal-line calculation unit 73 obtains a combination of luminance values at each position from the input three images 11a to 11c, and the normal-luminance table 72 corresponding to the camera that captured the image is Reference is made to determine the orientation of the normal at that position.
- the coordinate conversion unit 8 represents the direction of the normal calculated from the image captured by the cameras 1 and 2 by the coordinate conversion process in a unified coordinate system. Since the directions of the normals acquired from the images taken by the cameras 1 and 2 are expressed in the respective camera coordinate systems, an error occurs if they are compared as they are. In particular, when the difference in the imaging direction of the camera is large, this error becomes large.
- the coordinate system is unified by converting the direction of the normal acquired from the image of the camera 2 captured from obliquely above to the coordinate system of the camera 1.
- the direction of the normal obtained from the image of the camera 1 may be converted to the coordinate system of the camera 2 to unify the coordinate system, or the direction of the normal obtained from the images of the cameras 1 and 2 may be different
- the coordinate system may be unified by converting to the coordinate system of.
- x 1 and y 1 represent coordinates in the captured image of the camera 1
- x 2 and y 2 represent coordinates in the captured image of the camera 2.
- Equation 1 p a11, p a12, p a13, p a21, p a22, p a23 is the rotation matrix R 1, R 1_11, R 1_12 , R 1_13, R 1_21, R 1_22, as each equal to R 1_23
- the rotation matrix R 2 can be similarly obtained for the camera 2 as well.
- a rotation matrix R 21 for converting the coordinate system of the camera 2 into the coordinate system of the camera 1 can be obtained by R 2 ⁇ 1 ⁇ R 1 .
- the corresponding point calculation unit 9 calculates corresponding pixels from two normal images having a unified coordinate system. This processing is performed by obtaining a normal in the same direction as the normal of the pixel of interest in the normal image of the camera 1 from the normal image of the camera 2. The process performed by the corresponding point calculation unit 9 will be described with reference to the flowchart of FIG.
- the corresponding point calculation unit 9 acquires two normal images A and B in which the coordinate system is unified (S1).
- the normal image A acquired from the image of the camera 1 uses the one obtained from the surface shape calculation unit 7 as it is, and the coordinate conversion unit for the normal image B acquired from the image of the camera 2 At 8 the one converted to the coordinate system of the camera 1 is used.
- an arbitrary pixel in one normal image (here, the normal image A) is selected as a target point (target pixel) (S2).
- a contrast point is selected from the epipolar line of the other normal image (here, the normal image B) (S3).
- the similarity between the target point of the normal image A and the contrast point of the normal image B is calculated by the similarity evaluation function (S4).
- the degree of similarity is calculated using the directions of normals at pixels around the attention point and the contrast point.
- FIG. 10A shows an example of a search window used to calculate the degree of similarity.
- an area of 5 pixels ⁇ 5 pixels around the point of interest is used as a search window.
- the similarity between the attention point and the comparison point is calculated based on the matching rate of the direction of the normal in the entire search window. More specifically, the inner product of normal vectors between normal images A and B is calculated at each point in the search window by the following equation, and the similarity is calculated based on the sum (see FIG. 10B). .
- the point with the highest similarity is obtained from among the points, and the point is determined as the corresponding point in the normal image B corresponding to the target point of the normal image A ( S6).
- the triangulation unit 10 calculates depth information (distance) for each position of the measurement object 4. Since this process is a known technique, the detailed description is omitted.
- the corresponding point between two images is searched using the direction of the normal which is the physical feature of the measurement object, so the difference between the camera characteristics and the arrangement Three-dimensional measurement is possible without being affected by
- the corresponding point search process based on the color (brightness value) of the physical surface as in the prior art, the error becomes large when the target surface is a mirror surface, and accurate three-dimensional measurement is difficult, but the method according to the present embodiment It becomes possible to perform three-dimensional measurement with high accuracy even if it is a specular object.
- Modification 1 In the above embodiment, after the direction of the normal is calculated from the captured image of the camera 2 with reference to the normal-luminance table, processing is performed to adjust the coordinate system of the normal image to the coordinate system of the camera 1 by coordinate conversion. doing. However, other methods may be used if the coordinate system is finally unified. For example, conversion processing may be performed on the data of normals stored in the normal-brightness table corresponding to the camera 2 so that the coordinate system of the camera 2 matches the coordinate system of the camera 1. In this way, the calculation result of the direction of the normal by the surface shape calculation unit 7 with respect to the image of the camera 2 is represented by the coordinate system of the camera 1.
- Modification 2 In the above embodiment, three illumination devices 3a to 3c for emitting white light are used to sequentially turn on these to capture images, and the direction of the normal is calculated from the three images.
- any method can be adopted as long as it is a method of capturing an image and acquiring the direction of the normal.
- the luminous color of three illumination devices is respectively made into three colors of R, G, and B, these are irradiated simultaneously, and the same effect as the above is acquired only by one photography by acquiring the intensity of each component light. You can get
- the direction of the normal is used as a physical feature of the surface to be measured.
- corresponding points between stereo images are searched using spectral characteristics of the target.
- light sources with different spectral characteristics are sequentially irradiated to the measurement object from the same position.
- This can be realized, for example, by providing color filters having different spectral characteristics depending on places (angles) in front of a white light source as shown in FIG. 11 and rotating this filter.
- a simple spectral characteristic for each pixel can be calculated by observing the object with a color camera using such an illumination device and measuring the luminance value at which the value becomes the highest.
- association is performed using a spectral characteristic map for each pixel obtained by a plurality of cameras.
- the subsequent processing is the same as that of the first embodiment.
- a plurality of light sources for emitting light from different directions are disposed, and shooting is performed with a camera while sequentially turning on these light sources. Further, as in the first embodiment, a sample having a known shape such as a sphere and a known reflection characteristic is prepared. Here, a plurality of samples having different reflection characteristics are used, and the luminance value under each light source is held as case data for each sample.
- a plurality of light sources are sequentially turned on, and a combination of luminance values under each light source is acquired.
- the luminance values are combined and compared with the case data to calculate the corresponding reflection characteristic for each pixel.
- the pixels in the images captured by the plurality of cameras are associated with each other using the reflection characteristic map for each pixel obtained by the plurality of cameras.
- the subsequent processing is the same as that of the first embodiment.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
Description
〈全体概要〉
図1は本実施形態に係る3次元計測装置の概要を示す図である。図2は、本実施形態に係る3次元計測装置の機能ブロックを示す図である。図1に示すように、ステージ5に配置された計測対象物4を、2つのカメラ1,2によって撮影する。計測対象物4には、異なる方向から3つ照明装置3a~cから白色光が照射されており、これらの照明装置3a~cを順次照射して、カメラ1,2はそれぞれ3枚の画像を撮影する。撮影された画像はコンピュータ6に取り込まれ、画像処理されて3次元計測が行われる。
[カメラ配置]
図3は、カメラの配置を説明する図である。図3に示すように、カメラ1は鉛直方向から計測対象物4を撮影し、カメラ2は鉛直方向から40度ずれた方向から計測対象物4を撮影する。
図4は、照明装置3a~3cの配置を説明する図である。図4Aは、鉛直方向から見た図であり、照明装置3a~3cの方位角配置を示す図である。図4Bは横方向から見た図であり、照明装置3a~3cの天頂角配置を示す図である。図に示すように、3つの照明装置3a~3cは、方位角が互いに120度ずつ異なる方向から、また、天頂角が40度の方向から、計測対象物に光を照射する。
表面形状算出部7は、カメラ1,2によって撮影されたそれぞれ3枚の画像から、計測対象物の各位置における法線の向きを算出する機能部である。表面形状算出部7の、より詳細な機能ブロック図を図5に示す。図に示すように、表面形状算出部7は、画像入力部71,法線-輝度テーブル72および法線算出部73を有する。
座標変換部8は、座標変換処理によってカメラ1,2によって撮影された画像から算出された法線の向きを、統一された座標系で表す。カメラ1,2によって撮影された画像から取得された法線の向きは、それぞれのカメラ座標系で表されているので、そのままで比較すると誤差が生じる。特に、カメラの撮像方向の違いが大きいとこの誤差が大きくなる。
対応点算出部9は、座標系が統一された2つの法線画像から、対応する画素を算出する。この処理は、カメラ1の法線画像内の着目画素における法線と同じ向きの法線を、カメラ2の法線画像中から求めることで行われる。対応点算出部9が行う処理を、図9のフローチャートを参照しながら説明する。
以上のようにして、2つの画像における対応点が求められたら、三角測量部10によって計測対象物4の各位置について、奥行き情報(距離)を算出する。この処理は公知の技術であるので詳しい説明は省略する。
本実施形態に係る3次元計測装置では、計測対象物の物理特徴である法線の向きを利用して、2つの画像間での対応点を探索しているので、カメラの特性や配置の違いによる影響を受けずに3次元計測が可能となる。従来のような物理表面の色(輝度値)に基づく対応点探索処理では対象表面が鏡面である場合には誤差が大きくなり、精度の良い3次元計測が困難であるが、本実施形態による方法を使えば鏡面物体であっても精度良く3次元計測を行うことが可能となる。
上記の実施形態では、カメラ2の撮影画像から法線-輝度テーブルを参照して法線の向きを算出した後、座標変換によって法線画像の座標系をカメラ1の座標系に合わせる処理を実行している。しかしながら、最終的に座標系の統一が取れればその他の方法によっても構わない。例えば、カメラ2に対応する法線-輝度テーブルに格納される法線のデータに対して、カメラ2の座標系をカメラ1の座標系に合わせる変換処理を施しても良い。このようにすれば、カメラ2の画像に対する表面形状算出部7による法線の向きの算出結果が、カメラ1の座標系で表されたものとなる。
上記の実施形態では、白色光を照射する3つの照明装置3a~3cを利用し、これらを順次点灯して画像を撮影し、3枚の画像から法線の向きを算出している。しかしながら、画像を撮影して法線の向きを取得する方法であれば、任意の方法を採用することができる。例えば、3つの照明装置の発光色をそれぞれR,G,Bの3色とし、これらを同時に照射して、各成分光の強度を取得することで、1回の撮影のみで上記と同様の効果を得ることができる。
第1の実施形態では、計測対象表面の物理特徴として法線の向きを利用したが、本実施形態では対象の分光特性を利用して、ステレオ画像間の対応点を探索する。
本実施形態では、計測対象表面の物理特徴として反射特性を利用して、ステレオ画像間の対応点を探索する。
3a,3b,3c 照明装置
4 計測対象物
6 コンピュータ
7 表面形状算出部
71 画像入力部
72 法線-輝度テーブル
73 法線算出部
8 座標変換部
9 対応点算出部
10 三角測量部
Claims (10)
- 鏡面物体である計測対象物の3次元形状を計測する3次元計測装置であって、
複数のカメラと、
前記複数のカメラによって撮影された画像のそれぞれから、前記計測対象物の表面の物理特徴を取得する特徴取得手段と、
前記物理特徴を用いて、前記複数のカメラによって撮影された画像間で対応する画素を探索する対応画素探索手段と、
対応する画素間の視差に基づいて、3次元測量を行う測量手段と、
を備えることを特徴とする3次元計測装置。 - 前記特徴取得手段が取得する前記計測対象物の表面の物理特徴は、表面の法線の向きである
ことを特徴とする請求項1に記載の3次元計測装置。 - 複数のカメラによって撮影される画像の座標系を、変換パラメータを用いて共通の座標系に変換する座標変換手段をさらに有し、
前記対応画素探索手段は、座標変換手段によって共通の座標系に変換された法線の向きを利用して、画像間で対応する画素を探索する
ことを特徴とする請求項2に記載の3次元計測装置。 - 前記座標変換手段における変換パラメータは、あらかじめ行われたカメラキャリブレーションによって得られたパラメータから抽出されたものである
ことを特徴とする請求項3に記載の3次元計測装置。 - 前記対応画素探索手段は、着目する画素を含む所定の広さの領域における物理特徴を比較して、画像間で対応する画素を探索する
ことを特徴とする請求項1~4のいずれかに記載の3次元計測装置。 - 鏡面物体である計測対象物の3次元形状を計測する3次元計測方法であって、
複数のカメラによって撮影された画像のそれぞれから、前記計測対象物の表面の物理特徴を取得する特徴取得工程と、
前記物理特徴を用いて、前記複数のカメラによって撮影された画像間で対応する画素を探索する対応画素探索工程と、
対応する画素間の視差に基づいて、3次元測量を行う測量工程と、
を含むことを特徴とする3次元計測方法。 - 前記特徴取得工程において取得される前記計測対象物の表面の物理特徴は、表面の法線の向きである
ことを特徴とする請求項6に記載の3次元計測方法。 - 複数のカメラによって撮影される画像の座標系を、変換パラメータを用いて共通の座標系に変換する座標変換工程をさらに含み、
前記対応画素探索工程では、座標変換工程において共通の座標系に変換された法線の向きを利用して、画像間で対応する画素を探索する
ことを特徴とする請求項7に記載の3次元計測方法。 - 前記座標変換工程で利用される変換パラメータは、あらかじめ行われたカメラキャリブレーションによって得られたパラメータから抽出されたものである
ことを特徴とする請求項8に記載の3次元計測方法。 - 前記対応画素探索工程では、着目する画素を含む所定の広さの領域における物理特徴を比較して、画像間で対応する画素を探索する
ことを特徴とする請求項6~9のいずれかに記載の3次元計測方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP09814636A EP2339292A1 (en) | 2008-09-18 | 2009-09-17 | Three-dimensional measurement apparatus and method thereof |
CN200980136902.5A CN102159917A (zh) | 2008-09-18 | 2009-09-17 | 三维计测装置及其方法 |
US13/119,824 US20110228052A1 (en) | 2008-09-18 | 2009-09-17 | Three-dimensional measurement apparatus and method |
KR1020117007085A KR101194936B1 (ko) | 2008-09-18 | 2009-09-17 | 3차원 계측 장치 및 그 방법 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008239114A JP2010071782A (ja) | 2008-09-18 | 2008-09-18 | 3次元計測装置およびその方法 |
JP2008-239114 | 2008-09-18 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010032792A1 true WO2010032792A1 (ja) | 2010-03-25 |
Family
ID=42039614
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2009/066272 WO2010032792A1 (ja) | 2008-09-18 | 2009-09-17 | 3次元計測装置およびその方法 |
Country Status (6)
Country | Link |
---|---|
US (1) | US20110228052A1 (ja) |
EP (1) | EP2339292A1 (ja) |
JP (1) | JP2010071782A (ja) |
KR (1) | KR101194936B1 (ja) |
CN (1) | CN102159917A (ja) |
WO (1) | WO2010032792A1 (ja) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011215102A (ja) * | 2010-04-02 | 2011-10-27 | Seiko Epson Corp | 光学式位置検出装置 |
JP2011215099A (ja) * | 2010-04-02 | 2011-10-27 | Seiko Epson Corp | 光学式位置検出装置 |
JP2011215100A (ja) * | 2010-04-02 | 2011-10-27 | Seiko Epson Corp | 光学式位置検出装置 |
JP2011215101A (ja) * | 2010-04-02 | 2011-10-27 | Seiko Epson Corp | 光学式位置検出装置 |
Families Citing this family (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8687046B2 (en) * | 2009-11-06 | 2014-04-01 | Sony Corporation | Three-dimensional (3D) video for two-dimensional (2D) video messenger applications |
JP5563372B2 (ja) * | 2010-05-20 | 2014-07-30 | 第一実業ビスウィル株式会社 | 外観検査装置 |
US8334985B2 (en) * | 2010-10-08 | 2012-12-18 | Omron Corporation | Shape measuring apparatus and shape measuring method |
JP5624457B2 (ja) * | 2010-12-28 | 2014-11-12 | 株式会社東芝 | 三次元データ処理装置、方法及びプログラム |
JP5365645B2 (ja) * | 2011-01-17 | 2013-12-11 | オムロン株式会社 | 基板検査装置および基板検査システムならびに基板検査結果の確認用画面の表示方法 |
JP5832278B2 (ja) | 2011-12-26 | 2015-12-16 | 三菱重工業株式会社 | カメラ計測システムのキャリブレーション方法 |
JP5861462B2 (ja) | 2012-01-17 | 2016-02-16 | オムロン株式会社 | はんだ検査のための検査基準登録方法およびその方法を用いた基板検査装置 |
WO2015132981A1 (ja) | 2014-03-03 | 2015-09-11 | 三菱電機株式会社 | 位置測定装置及び位置測定方法 |
JP6432968B2 (ja) * | 2014-06-26 | 2018-12-05 | 国立大学法人岐阜大学 | 物体形状推定装置及びプログラム |
US9818021B2 (en) * | 2014-10-21 | 2017-11-14 | Isra Surface Vision Gmbh | Method for determining a local refractive power and device therefor |
DE102014115331A1 (de) * | 2014-10-21 | 2016-04-21 | Isra Surface Vision Gmbh | Verfahren und Vorrichtung zur Ermittlung einer dreidimensionalen Verzerrung |
CN107548449B (zh) * | 2015-04-21 | 2019-11-12 | 卡尔蔡司工业测量技术有限公司 | 用于确定被测对象的实际尺寸特征的方法和装置 |
US20160378137A1 (en) * | 2015-06-26 | 2016-12-29 | Intel Corporation | Electronic device with combinable image input devices |
WO2017052590A1 (en) | 2015-09-25 | 2017-03-30 | Intel Corporation | Hinge for an electronic device |
JP6671915B2 (ja) * | 2015-10-14 | 2020-03-25 | キヤノン株式会社 | 処理装置、処理システム、撮像装置、処理方法、プログラム、および記録媒体 |
CN106524909B (zh) * | 2016-10-20 | 2020-10-16 | 北京旷视科技有限公司 | 三维图像采集方法及装置 |
JP6848385B2 (ja) * | 2016-11-18 | 2021-03-24 | オムロン株式会社 | 三次元形状計測装置 |
US20190096057A1 (en) | 2017-05-11 | 2019-03-28 | Jacob Nathaniel Allen | Object inspection system and method for inspecting an object |
JPWO2019124104A1 (ja) * | 2017-12-20 | 2020-12-17 | ソニー株式会社 | 物体形状計測装置、および方法、並びにプログラム |
EP4269940A3 (en) * | 2018-10-15 | 2023-11-29 | Koh Young Technology Inc. | Device and method for inspection |
JP6858415B2 (ja) * | 2019-01-11 | 2021-04-14 | 学校法人福岡工業大学 | 海面計測システム、海面計測方法および海面計測プログラム |
CN111915666A (zh) * | 2019-05-07 | 2020-11-10 | 顺丰科技有限公司 | 基于移动终端的体积测量方法及装置 |
JP7242431B2 (ja) * | 2019-05-31 | 2023-03-20 | 公益財団法人かずさDna研究所 | 三次元計測装置、三次元計測方法および三次元計測用プログラム |
JP7452091B2 (ja) | 2020-02-27 | 2024-03-19 | オムロン株式会社 | X線検査システム、x線検査方法及びプログラム |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS61198015A (ja) | 1984-11-14 | 1986-09-02 | Agency Of Ind Science & Technol | 二組の照度差ステレオ法にもとづく距離計測法及びその装置 |
JPH04143606A (ja) * | 1990-10-04 | 1992-05-18 | Kobe Steel Ltd | 形状検出装置 |
JP2007114168A (ja) * | 2005-10-17 | 2007-05-10 | Applied Vision Systems Corp | 画像処理方法および装置、並びにプログラム |
JP2007322162A (ja) * | 2006-05-30 | 2007-12-13 | 3D Media Co Ltd | 3次元形状測定装置及び3次元形状測定方法 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS6282314A (ja) * | 1985-10-08 | 1987-04-15 | Hitachi Ltd | 光度差ステレオ計測方式 |
US20020024517A1 (en) * | 2000-07-14 | 2002-02-28 | Komatsu Ltd. | Apparatus and method for three-dimensional image production and presenting real objects in virtual three-dimensional space |
-
2008
- 2008-09-18 JP JP2008239114A patent/JP2010071782A/ja active Pending
-
2009
- 2009-09-17 CN CN200980136902.5A patent/CN102159917A/zh active Pending
- 2009-09-17 US US13/119,824 patent/US20110228052A1/en not_active Abandoned
- 2009-09-17 KR KR1020117007085A patent/KR101194936B1/ko not_active IP Right Cessation
- 2009-09-17 WO PCT/JP2009/066272 patent/WO2010032792A1/ja active Application Filing
- 2009-09-17 EP EP09814636A patent/EP2339292A1/en not_active Withdrawn
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS61198015A (ja) | 1984-11-14 | 1986-09-02 | Agency Of Ind Science & Technol | 二組の照度差ステレオ法にもとづく距離計測法及びその装置 |
JPH04143606A (ja) * | 1990-10-04 | 1992-05-18 | Kobe Steel Ltd | 形状検出装置 |
JP2007114168A (ja) * | 2005-10-17 | 2007-05-10 | Applied Vision Systems Corp | 画像処理方法および装置、並びにプログラム |
JP2007322162A (ja) * | 2006-05-30 | 2007-12-13 | 3D Media Co Ltd | 3次元形状測定装置及び3次元形状測定方法 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2011215102A (ja) * | 2010-04-02 | 2011-10-27 | Seiko Epson Corp | 光学式位置検出装置 |
JP2011215099A (ja) * | 2010-04-02 | 2011-10-27 | Seiko Epson Corp | 光学式位置検出装置 |
JP2011215100A (ja) * | 2010-04-02 | 2011-10-27 | Seiko Epson Corp | 光学式位置検出装置 |
JP2011215101A (ja) * | 2010-04-02 | 2011-10-27 | Seiko Epson Corp | 光学式位置検出装置 |
Also Published As
Publication number | Publication date |
---|---|
EP2339292A1 (en) | 2011-06-29 |
CN102159917A (zh) | 2011-08-17 |
US20110228052A1 (en) | 2011-09-22 |
KR101194936B1 (ko) | 2012-10-25 |
JP2010071782A (ja) | 2010-04-02 |
KR20110059631A (ko) | 2011-06-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2010032792A1 (ja) | 3次元計測装置およびその方法 | |
US9392262B2 (en) | System and method for 3D reconstruction using multiple multi-channel cameras | |
US8326022B2 (en) | Stereoscopic measurement system and method | |
US8345953B2 (en) | Stereoscopic measurement system and method | |
US9449378B2 (en) | System and method for processing stereoscopic vehicle information | |
EP2870428A1 (en) | 3-d scanning and positioning system | |
EP3382645A2 (en) | Method for generation of a 3d model based on structure from motion and photometric stereo of 2d sparse images | |
US9286506B2 (en) | Stereoscopic measurement system and method | |
CA2757313C (en) | Stereoscopic measurement system and method | |
JP3221384B2 (ja) | 三次元座標計測装置 | |
CA2757323C (en) | Stereoscopic measurement system and method | |
CA2757321C (en) | Stereoscopic measurement system and method | |
EP3688407A1 (en) | Light projection systems | |
JP2022177166A (ja) | 検査方法、プログラム、及び、検査システム | |
AU2013292286A1 (en) | System and method for processing stereoscopic vehicle information |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200980136902.5 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09814636 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009814636 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 20117007085 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13119824 Country of ref document: US |