CN208653473U - Image capture device, 3D information comparison device, mating object generating means - Google Patents

Image capture device, 3D information comparison device, mating object generating means Download PDF

Info

Publication number
CN208653473U
CN208653473U CN201821449456.6U CN201821449456U CN208653473U CN 208653473 U CN208653473 U CN 208653473U CN 201821449456 U CN201821449456 U CN 201821449456U CN 208653473 U CN208653473 U CN 208653473U
Authority
CN
China
Prior art keywords
video data
video
image capture
image
capture device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201821449456.6U
Other languages
Chinese (zh)
Inventor
左忠斌
左达宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianmu Love Vision (beijing) Technology Co Ltd
Tianmu Aishi Beijing Technology Co Ltd
Original Assignee
Tianmu Love Vision (beijing) Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianmu Love Vision (beijing) Technology Co Ltd filed Critical Tianmu Love Vision (beijing) Technology Co Ltd
Priority to CN201821449456.6U priority Critical patent/CN208653473U/en
Application granted granted Critical
Publication of CN208653473U publication Critical patent/CN208653473U/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

The utility model provides a kind of image capture device, 3D information comparison device, mating object generating means, wherein image capture device includes: that video data provides device, for providing the video data including object;Video process apparatus, for handling the video data, to obtain multiple images from the video data;Processing unit, for obtaining the 3D information of object according at least three in multiple images;Measuring device, according to the geometric dimension of the 3D information measurement object of object.It notices for the first time and proposes to synthesize the occasion that 3D is not particularly suited for certain movabilities, easily-deformable object and image capture device have been fixed in such a way that camera shoots picture.And propose to carry out 3D synthesis using video data, acquisition speed is improved, object interior movability and deformation when shooting are avoided.Meanwhile the hardware resource of video acquisition is made full use of, it avoids repeating to install the raising of bring cost.

Description

Image capture device, 3D information comparison device, mating object generating means
Technical field
The utility model relates to object field of measuring technique, in particular to carry out object 3D acquisition and length using video Etc. geometric dimensions field of measuring technique.
Background technique
3D acquisition at present/measuring device mainly utilizes camera around object rotation acquisition object multiple pictures, to synthesize The 3D rendering of object, and the measurement such as object length, profile is carried out using 3D point cloud data.In this process, it is desirable that object appearance State is relatively fixed.If in shooting process, object is mobile, or deformation occurs, then the multiple pictures that will lead to shooting are difficult to synthesize 3D rendering.This requirement is easier realization in general.
However, for example carrying out the 3D acquisition and measurement of iris, micro-lens shooting is generallyd use, camera is shooting different photographs Need to shoot around iris in different location when piece, but the variation of position cause micro-lens to change position every time will be right manually It is burnt.It is elongated to will lead to entire shooting time in this way.But in this process, since the time is longer, eyes are easy to rotate, and cause The plurality of pictures of shooting is not under the same posture of iris, to influence 3D synthesis.That is, shooting some movabilities, When easily-deformable object, traditional approach is difficult to accurate 3D synthesis, to affect synthetic effect and measurement accuracy.
In addition, many occasion video cameras are more universal at present, for example, in access control system, in safety check monitoring, in bank ATM Monitored in automatic teller machine, in traffic offence etc., it is all made of video camera and records object by way of video.Therefore, if it is desired to 3D acquisition, measurement, identification and comparison are carried out in these occasions then to need that camera additionally is installed, and cause larger cost.
Under normal conditions, mainly for fixed still life, the limitation in this application is hindered for 3D acquisition and measuring technique The discovery of above-mentioned technical problem does not more attempt the motivation for solving the technical problem.
Utility model content
In view of the above problems, the utility model is proposed to overcome the above problem in order to provide one kind or at least partly solve The certainly image capture device based on video data of the above problem.
The utility model provides a kind of image capture device based on video data, including
Video data provides device, for providing the video data including object;
Video process apparatus, for handling the video data, to obtain multiple images from the video data;
Processing unit, for obtaining the 3D information of object according at least three in multiple images;
Measuring device, for the geometric dimension according to the 3D information measurement object of the object.
The utility model additionally provides a kind of image capture device based on video data, including
Video data provides device, for providing the video data including object;
Video process apparatus, for handling the video data, to obtain multiple images from the video data.
The utility model additionally provides a kind of image capture device based on video data, including
Video data provides device, for providing the video data including object;
Video process apparatus, for handling the video data, to obtain multiple images from the video data;
Processing unit, for obtaining the 3D information of object according at least three in described multiple images.
Optionally, in described multiple images, the target object area of at least two images partially overlaps.
Optionally, it includes image collecting device that the video data, which provides device,.
Optionally, the video data is generated by the pickup area of image collecting device and the relative motion of object.
Optionally, it includes the direct interception image from video data that multiple images are obtained from the video data, or will Interception image after compressed video data restores.
Optionally, the video data provides device and provides one or more video datas.
Optionally, when video data, which provides device, provides multiple video datas, for obtaining the more of object 3D information A image comes from the same video data, or comes from different video data.
Optionally, the video frequency data frame rate is greater than 5 frames/second.
Optionally, the acquisition position of the two adjacent images in described multiple images at least meets following condition:
H* (1-cosb)=L*sin2b;
A=m*b;
0<m<0.8
Wherein L is distance of the image collecting device to object, and H is object actual size in acquired image, and a is Two neighboring location drawing picture acquisition device optical axis included angle, m are coefficient.
Optionally, adjacent three positions in described multiple images meet three images acquired on corresponding position at least There is the part for indicating object the same area.
The utility model additionally provides a kind of 3D information comparison device, including described in any one based on video counts According to image capture device.
The utility model additionally provides a kind of mating object generating means of object, utilizes base described in any one 3D information at least one region of the object that the image capture device of video data obtains generates and object respective area The mating object that domain matches.
Utility model inventive point and technical effect
1, it notices for the first time and proposes that synthesizing 3D in such a way that camera shoots picture is not particularly suited for certain movabilities, mutability The occasion that the object and image capture device of shape have been fixed.And propose to carry out 3D synthesis using video data, improve acquisition Speed avoids object interior movability and deformation when shooting.Meanwhile the hardware resource of video acquisition is made full use of, avoid weight Installation bring cost improves again.
2, according to repetition test, the optimized parameter range using video data synthesis 3D information is proposed, to improve Synthetic effect and measurement accuracy, and the motion conditions of object can be taken into account.
3, the prior art mainly passes through HardwareUpgring and stringent calibration for the promotion of synthetic effect, does not have in the prior art Angle position when any enlightenment can be taken pictures by changing video camera guarantees the effect and stability of 3D synthesis, more without specific The condition of optimization.Guarantee the effect and stabilization of 3D synthesis present invention firstly provides angle position when optimization camera photography Property, and by repetition test, the best practices condition that camera position needs to meet is proposed, 3D synthesis is substantially increased Effect and composograph stability.
Detailed description of the invention
By reading the following detailed description of the preferred embodiment, various other advantages and benefits are common for this field Technical staff will become clear.The drawings are only for the purpose of illustrating a preferred embodiment, and is not considered as practical to this Novel limitation.And throughout the drawings, the same reference numbers will be used to refer to the same parts.In the accompanying drawings:
Fig. 1 shows the schematic diagram of the image capture device based on video data according to an embodiment of the present invention;
Fig. 2 shows specific according to one kind of the image capture device based on video data of an embodiment of the present invention The schematic diagram of embodiment;
Fig. 3 shows another specific according to the image capture device based on video data of an embodiment of the present invention The schematic diagram of embodiment;
Fig. 4 shows a kind of specific according to the image capture device based on video data of another embodiment of the utility model The schematic diagram of embodiment;
Fig. 5 shows a kind of specific according to the image capture device based on video data of another embodiment of the utility model The schematic diagram of embodiment;
Description of symbols:
100 video datas provide device,
1000 video cameras,
200 video process apparatus,
300 image processing apparatus,
101 mobile devices,
102 rotating devices,
400 range units,
1011 tracks,
1012 adjustment devices.
Specific embodiment
Exemplary embodiments of the present disclosure are described in more detail below with reference to accompanying drawings.Although showing the disclosure in attached drawing Exemplary embodiment, it being understood, however, that may be realized in various forms the disclosure without should be by embodiments set forth here It is limited.On the contrary, these embodiments are provided to facilitate a more thoroughly understanding of the present invention, and can be by the scope of the present disclosure It is fully disclosed to those skilled in the art.
Embodiment 1
In order to solve the above technical problems, an embodiment of the utility model provides a kind of 3D information based on video data Acquisition/measuring device.As shown in Figure 1, specifically including: video data provides device 100, video process apparatus 200, image procossing Device 300.
Video data provides device 100, can be video camera, the camera with camera function, mobile phone, plate, monitoring Device etc. includes the video of object for the shooting of alignment target object, generates video data, and video data is sent at video Manage device 200.Above-mentioned video data provides device 100 and can arrange in pairs or groups as needed various camera lenses, such as infrared lens, visible light microscopic Head, remote zoom lens, wide-angle lens, Macro Lens head etc..Above-mentioned video data provide device 100 can also for memory, with depositing The processor etc. of function is stored up, storage inside has the video data including object.
By taking video camera as an example, in one case, as shown in Fig. 2, video camera 1000 is mounted in mobile device 101, move Dynamic device 101 can be moved along track 1011, to drive 1000 surrounding target object of video camera to rotate, so that shooting includes mesh Mark the video of object different angle.Mobile device 101 or mechanical arm drive 1000 surrounding target of video camera by mechanical arm Object rotation.
In one case, as shown in figure 3, video camera 1000 is mounted on rotating device 102, rotating device 102 is in Mandrel drives video camera 1000 to rotate, so that shooting includes the video of object different angle.
In one case, video camera 1000 is fixed, and object (such as human body, head) is mobile or rotates, so that camera shooting Machine 1000 can take the video comprising object different angle.
Video process apparatus 200 can be the processor comprising program, computer, mobile phone, plate etc., for receiving camera shooting The video data that machine 1000 is sent, and multiple images are obtained from above-mentioned video data.Processing method includes:
Step 1: reading video data;
Step 2: obtaining the frame number of video data.It can use the frame number that software command obtains video data, while can also Voluntarily to be provided by video data.
Step 3: according to frame number, framing being carried out to video data, circulation extracts every frame data;
Step 4: image data is recovered according to every frame video data.
It can be realized by some program codes by obtaining image in video, such as:
Particularly, the conversion of video-picture can also be carried out with the mode of video interception in some cases.In some feelings Under condition, image is recovered after video data can also being decompressed.
Image processing apparatus 300 obtains the 3D information of object for handling above-mentioned multiple images, i.e., according to above-mentioned multiple The overlapped part of image, all or part of by them are stitched together, and realize the synthesis of 3D information.The method of image mosaic Include:
(1) multiple images are handled, extracts respective characteristic point;The feature of respective characteristic point can in multiple images To be retouched using SIFT (Scale-Invariant Feature Transform, scale invariant feature conversion) Feature Descriptor It states.SIFT feature description has 128 feature description vectors, and the 128 of any characteristic point can be described on direction and scale The feature of a aspect significantly improves the precision to feature description, while Feature Descriptor has independence spatially.
(2) characteristic point of the multiple images based on extraction, feature point cloud data and the iris for generating face characteristic respectively are special The feature point cloud data of sign.It specifically includes:
(2-1) carries out the spy of plurality of pictures according to the feature of the respective characteristic point of each image in the multiple images of extraction The matching for levying point, establishes matched facial feature points data set;According to the respective feature of each image in the multiple images of extraction The feature of point, carries out the matching of the characteristic point of plurality of pictures, establishes matched iris feature point data collection;
(2-2) according to the optical information of camera, obtain multiple images when camera different location, calculate each position phase Relative position of the machine relative to characteristic point spatially, and the space of the characteristic point in multiple images is calculated depending on the relative position Depth information.Similarly, the spatial depth information of the characteristic point in multiple images can be calculated.Bundle adjustment can be used in calculating Method.
The spatial depth information for calculating characteristic point may include: spatial positional information and colouring information, that is, can be feature Point is in the X axis coordinate of spatial position, characteristic point in the Y axis coordinate of spatial position, characteristic point in the Z axis coordinate of spatial position, spy The B for levying the colouring information of the value in the channel R of the colouring information of point, the value in the channel G of the colouring information of characteristic point, characteristic point is logical The value in road, the value in the channel Alpha of the colouring information of characteristic point etc..In this way, containing spy in the feature point cloud data generated The spatial positional information and colouring information of point are levied, the format of feature point cloud data can be as follows:
X1Y1Z1R1G1B1A1
X2Y2Z2R2G2B2A2
……
Xn Yn Zn Rn Gn Bn An
Wherein, Xn indicates characteristic point in the X axis coordinate of spatial position;Yn indicates characteristic point in the Y axis coordinate of spatial position; Zn indicates characteristic point in the Z axis coordinate of spatial position;Rn indicates the value in the channel R of the colouring information of characteristic point;Gn indicates feature The value in the channel G of the colouring information of point;Bn indicates the value of the channel B of the colouring information of characteristic point;The color of An expression characteristic point The value in the channel Alpha of information.
(2-3) generates object according to the spatial depth information of multiple images matched characteristic point data collection and characteristic point The feature point cloud data of feature.
(2-4) constructs object 3D model according to feature point cloud data, to realize the acquisition of object point cloud data.
Collected object color, texture are attached on point cloud data by (2-5), form object 3D image.
Wherein it is possible to 3D rendering is synthesized using all images in one group of image, it can also be higher from wherein selection quality Image synthesized.
Above-mentioned joining method is limited citing, however it is not limited to which this, several with good grounds multi-angle two dimensional images of institute generate three The method of dimension image can be used.
According to experiment, the frame rate of the video can take into account synthesis quality in 5-500 frame/second, and can be to movement Object accurately synthesized.Such as when shooting iris, human eye can rotate in spite of oneself, if frame per second is too low, repeatedly clap The iris position arrived is not consistent, influences synthetic effect, therefore the utility model test of many times, according to human eye movement's feature, visits Rope goes out the zone of reasonableness of the parameter, this is also one of inventive point of the utility model.
Embodiment 2
In one case, as shown in figure 4, video data provide device 100 include multiple video camera 100A, 100B, 100C ... 100N, each video camera distinguish the video A of photographic subjects object, video B, video C ... video N from different directions.
Video process apparatus 200 extracts image A1, A2 ... An from video A respectively;From video B extract image B1, B2…Bn;Image C1, C2 ... Cn are extracted from video C;... extracts image N1, N2 ... Nn from video N;Wherein N and n all generations Table can according to need the number constantly increased.
Image processing apparatus 300 utilizes any some progress 3D figures in An, Bn, Cn ... .Nn in above-mentioned n group image Piece synthesis.Any one picture that can choose in each video is synthesized, and also can choose the part figure in partial video Piece is synthesized, that is to say, that from above-mentioned picture set, all or part of any image from different video of selection into Row synthesis.Certainly, these pictures of selection should meet the coincidence of Partial Feature point.
Under special circumstances, object can deformation or movement, then multiple video camera 100A, 100B, 100C ... 100N respectively the video A of photographic subjects object, video B, video C ... video N from different directions when, should also records photographing open Begin the information such as time, frame per second.Video process apparatus 200 is extracting image A1, A2 ... An from video A respectively;From video B Extract image B1, B2 ... Bn;Image C1, C2 ... Cn are extracted from video C;... extracts image N1, N2 ... Nn's from video N Meanwhile stamping timestamp for every image, i.e., absolute time when video camera collects the image content you that.Image procossing Device 300 carries out 3D synthesis using the consistent image of timestamp in above-mentioned n group image, that is, guarantees multiple figures for 3D synthesis Piece information is absolute synchronization shooting.That is, even if the object of which movement in entire shooting process, each video camera can also be with In the image information of the synchronization acquisition object moment all directions, to guarantee that synthesis is accurate.
Particularly, in this case, image processing apparatus 300 can synthesize multiple moment objects 3D image or 3D model, that is to say, that available object when shooting grow in a series of 3D renderings or 3D point cloud model.If even Available object 3D rendering is put in continued broadcasting or 3D point cloud model changes with time.In this way, can restore in security protection application The movement posture of object 3D rendering or 3D point cloud model out, such as the posture of monitored someone's walking can be recovered, And this posture, it can be shown with 3D, it can watch the posture that the people walks from any angle.In this way, for monitoring mesh Target matching identification is not limited only to the information such as profile, characteristic point, surface texture, can also be according to the 3D motion posture of monitoring objective It carries out.This is also one of inventive point of the utility model.Certainly, such processing can bring the significantly increasing of data processing amount Add, but hardware computing capability can be improved in the modes such as parallel computation, cloud computing, realizes above-mentioned target.
Embodiment 3
In above-mentioned multiple embodiments, video camera 1000 can using video camera it is general focus in real time by the way of shoot. Meanwhile as shown in figure 5, in a certain position Location1, the measurement of range unit 400 is imaged in order to improve focusing speed The a certain face region distance of machine 1000 and object is H1, and lens focus is F1 at this time;When video camera 1000 is located at another one When setting Location12 or object and having rotated certain angle relative to video camera 1000, the measurement of range unit 400 is imaged Another face region distance of machine 1000 and object is H2, and controller receives the data that range unit 400 is sent, and according to object Away from-lens focus table (measurement obtains in advance) judgement focal length F2 that camera lens should have at this time, lens motor is controlled, so as to adjust Lens focus is to F2.
Wherein range unit 400 can be the various types such as laser range finder, image rangefinder.It can individually become one A module is also possible to a part of video camera 1000.
Other than distance measuring method, in position A2, it is Q1 that video camera 1000, which collects picture contrast, at this time processing unit 100 control lens motors rotations, adjustment lens focus becomes larger or becomes smaller so that after reaching a certain contrast Q2, no matter focal length again Secondary increase or reduction, Q2 reduce, that is to say, that the contrast of Q2 is best.Focal length F2 is pinpointed focus at this time, with the focal length Shoot this region apparent.
It is of course also possible to judge to readjust focal length by the collected object size variation of video camera 1000, make Ratio of the object in 1000 visual field of video camera is obtained to remain unchanged.
The acquisition position of two adjacent images in above-mentioned multiple images at least meets following condition:
H* (1-cosb)=L*sin2b;
A=m*b;
0<m<0.8
Wherein L is distance of the image collecting device to object, and H is object actual size in acquired image, and a is Two neighboring location drawing picture acquisition device optical axis included angle, m are coefficient.
For one video camera, H is object actual size in acquired image, and image is usually video camera at first The picture shot when setting, the object in the picture have true geometric dimension (not being the size in picture), measure the ruler Along the orientation measurement of first position to the second position when very little.Such as first position and the second position are the relationships moved horizontally, So the size is measured along the horizontal cross of object.Such as the object left end that can show that in picture is A, it is most right End is B, then measures the linear distance of A to B on object, is H.The measurement method can be tied according to A, B distance in picture It closes camera lens focal length and carries out actual distance calculation, A, B can also be identified on object, it is straight using other measurement means Connect measurement AB linear distance.
A is two neighboring location drawing picture acquisition device optical axis included angle.
M is coefficient
Since article size, concave-convex situation are different, the value of a can not be limited with strict formula, needs rule of thumb to carry out It limits.According to many experiments, the value of m preferably can be within 0.8 within 1.5.Specific experiment data referring to Following table:
Object M value Synthetic effect Synthetic ratio
Human body head 0.1、0.2、0.3、0.4 It is very good > 90%
Human body head 0.5、0.6 It is good > 85%
Human body head 0.7、0.8 It is relatively good > 80%
Human body head 0.9、1.0 Generally > 70%
Human body head 1.0、1.1、1.2 Generally > 60%
Human body head 1.2、1.3、1.4、1.5 Synthesis reluctantly > 50%
Human body head 1.6、1.7 It is difficult to synthesize < 40%
After object and video camera determine, the value of a can be calculated according to above-mentioned empirical equation, it can be true according to a value Determine the parameter of virtual matrix, i.e. positional relationship between matrix dot.
In general, virtual matrix is one-dimensional matrix, such as along horizontal direction arrangement multiple matrix dots (acquisition Position).But when some target objects are larger, two-dimensional matrix is needed, then two adjacent in vertical direction positions are equally full The above-mentioned a value condition of foot.
Under some cases, even from above-mentioned empirical equation, also it is not easy to determine a value under some occasions, needs basis at this time Experiment adjustment matrix parameter, experimental method are as follows: calculating prediction matrix parameter a according to above-mentioned formula, and according to matrix parameter control Video camera processed is moved to corresponding matrix dot, such as video camera shoots picture P1 in position W1, is moved to shooting figure after the W2 of position Piece P2, compare at this time whether have in picture P1 and picture P2 indicate object the same area part, i.e., P1 ∩ P2 non-empty (such as Simultaneously include human eye angle part, but photograph taking angle is different), if readjusting a value without if, re-move to position W2 ' repeats above-mentioned comparison step.If P1 ∩ P2 non-empty, video camera is continued to move to extremely according to a value (adjustment or unadjusted) The position W3 shoots picture P3, and whether compare to have in picture P1, picture P2 and picture P3 again indicates object the same area Part, i.e. P1 ∩ P2 ∩ P3 non-empty.It recycles plurality of pictures to synthesize 3D, tests 3D synthetic effect, meet 3D information collection and survey Amount requires.That is, the structure of matrix is determined by the position of video camera when acquisition multiple images, adjacent three positions It sets three images for meeting and acquiring on corresponding position and at least there is the part for indicating object the same area.
Embodiment 4
In some cases, after 1000 run-home object of video camera, size accounting is lower in the picture for object, needs Further camera lens, the resolution ratio of object when just can guarantee 3D synthesis in this way.
Therefore, it please refers to shown in Fig. 5, further includes adjustment device 1012, size accounting drives in the picture according to object Device 1012 is adjusted, so that it drives video camera 1000 to the direction movement close to object or separate object.Adjust device 1012 between mobile device 101 and video camera 1000.Adjusting device 1012 can be an axis motion platform, two axle movement Platform, triaxial movement platform, four axes motion platform, 5-axis movement platform or six axis motion platforms.
Video camera 1000 includes lens driving motor, according to object size accounting drive motor in the picture, band index glass Head is mobile, so that object size accounting in the picture is increased or reduced, so that ratio is appropriate.Above and this is practical new The inventive point of type.
Object described in the utility model can be a physical objects, or multiple objects constituent.
The 3D information of the object includes 3D rendering, 3D point cloud, 3D grid, part 3D feature, 3D size and all bands There is the parameter of object 3D feature.
So-called 3D, three-dimensional refer to tri- directional informations of XYZ in the utility model, especially have depth information, There is essential distinction with only two-dimensional surface information.Also it is known as 3D, panorama, holography, three-dimensional with some, but actually only includes two Information is tieed up, does not especially include that the definition of depth information has essential distinction.
Pickup area described in the utility model refers to the range that image collecting device (such as camera) can be shot.
Image collecting device in the utility model can for CCD, CMOS, camera, video camera, industrial camera, monitor, Camera, mobile phone, plate, notebook, mobile terminal, wearable device, smart glasses, smart watches, Intelligent bracelet and band There are image collecting function all devices.
The 3D information for the object multiple regions that above embodiments obtain can be used for being compared, such as identity Identification.The 3D information of human body face and iris is obtained first with the embodiment of the utility model, and is stored it in server, As normal data.When in use, being operated such as needing to carry out authentication and paid, opened the door, it can be obtained with 3D Device acquires again and obtains the 3D information of human body face and iris, it is compared with normal data, compares and successfully then permits Perhaps next step movement is carried out.It is appreciated that this compare the identification that can be used for the fixtures such as antique, the art work, i.e., first Obtain antique, art work multiple regions 3D information as normal data, when needing to identify, again acquisition multiple regions 3D Information, and be compared with normal data, it discerns the false from the genuine.
The 3D information for the object multiple regions that above embodiments obtain can be used for designing for the object, production, make Make mating object.For example, obtaining human body head 3D data, it can be human design, manufacture more particularly suitable cap;Obtain human body head Portion's data and eyes 3D data can be human design, the suitable glasses of manufacture.
Above embodiments obtain object 3D information can be used for the geometric dimension to the object, appearance profile into Row measurement.
In the instructions provided here, numerous specific details are set forth.It is to be appreciated, however, that the utility model Embodiment can be practiced without these specific details.In some instances, be not been shown in detail well known method, Structure and technology, so as not to obscure the understanding of this specification.
Similarly, it should be understood that in order to simplify the disclosure and help to understand one or more of the various inventive aspects, Above in the description of exemplary embodiment of the present invention, each feature of the invention is grouped together into single implementation sometimes In example, figure or descriptions thereof.However, the disclosed method should not be interpreted as reflecting the following intention: i.e. required to protect Shield the present invention claims features more more than feature expressly recited in each claim.More precisely, as following Claims reflect as, inventive aspect is all features less than single embodiment disclosed above.Therefore, Thus the claims for following specific embodiment are expressly incorporated in the specific embodiment, wherein each claim itself All as a separate embodiment of the present invention.
Those skilled in the art will understand that can be carried out adaptively to the module in the equipment in embodiment Change and they are arranged in one or more devices different from this embodiment.It can be the module or list in embodiment Member or component are combined into a module or unit or component, and furthermore they can be divided into multiple submodule or subelement or Sub-component.Other than such feature and/or at least some of process or unit exclude each other, it can use any Combination is to all features disclosed in this specification (including adjoint claim, abstract and attached drawing) and so disclosed All process or units of what method or apparatus are combined.Unless expressly stated otherwise, this specification is (including adjoint power Benefit require, abstract and attached drawing) disclosed in each feature can carry out generation with an alternative feature that provides the same, equivalent, or similar purpose It replaces.
In addition, it will be appreciated by those of skill in the art that although some embodiments in this include institute in other embodiments Including certain features rather than other feature, but the combination of the feature of different embodiment means in the scope of the present invention Within and form different embodiments.For example, in detail in the claims, the one of any of embodiment claimed all may be used Come in a manner of in any combination using.
Various component embodiments of the invention can be implemented in hardware, or to run on one or more processors Software module realize, or be implemented in a combination thereof.It will be understood by those of skill in the art that can be used in practice Microprocessor or digital signal processor (DSP) realize some or all portions in device according to an embodiment of the present invention The some or all functions of part.The present invention is also implemented as a part or complete for executing method as described herein The device or device program (for example, computer program and computer program product) in portion.It is such to realize program of the invention It can store on a computer-readable medium, or may be in the form of one or more signals.Such signal can be with It downloads from internet website, is perhaps provided on the carrier signal or is provided in any other form.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and ability Field technique personnel can be designed alternative embodiment without departing from the scope of the appended claims.In the claims, Any reference symbol between parentheses should not be configured to limitations on claims.Word "comprising" does not exclude the presence of not Element or step listed in the claims.Word "a" or "an" located in front of the element does not exclude the presence of multiple such Element.The present invention can be by means of including the hardware of several different elements and being come by means of properly programmed computer real It is existing.In the unit claims listing several devices, several in these devices can be through the same hardware branch To embody.The use of word first, second, and third does not indicate any sequence.These words can be explained and be run after fame Claim.
So far, although those skilled in the art will appreciate that present invention has been shown and described in detail herein multiple shows Example property embodiment still without departing from the spirit and scope of the present invention, still can according to the present disclosure directly Determine or deduce out many other variations or modifications consistent with the principles of the invention.Therefore, the scope of the present invention is understood that and recognizes It is set to and covers all such other variations or modifications.

Claims (14)

1. a kind of image capture device, it is characterised in that: including
Video data provides device, for providing the video data including object;
Video process apparatus, for handling the video data, to obtain multiple images from the video data;
Processing unit, for obtaining the 3D information of object according at least three in multiple images;
Measuring device, for the geometric dimension according to the 3D information measurement object of the object.
2. a kind of image capture device, it is characterised in that: including
Video data provides device, for providing the video data including object;
Video process apparatus, for handling the video data, to obtain multiple images from the video data.
3. image capture device as claimed in claim 2, it is characterised in that: in described multiple images, at least two images Target object area partially overlap.
4. image capture device as claimed in claim 2, it is characterised in that: it includes that image is adopted that the video data, which provides device, Acquisition means.
5. image capture device as claimed in claim 4, it is characterised in that: the video data is adopted by image collecting device The relative motion for collecting region and object generates.
6. image capture device as claimed in claim 2, it is characterised in that: obtain multiple images packet from the video data The direct interception image from video data is included, or interception image after compressed video data is restored.
7. image capture device as claimed in claim 2, it is characterised in that: the video data provide device provide one or Multiple video datas.
8. image capture device as claimed in claim 7, it is characterised in that: provide multiple videos when video data provides device When data, the multiple images for obtaining object 3D information come from the same video data, or come from different video data.
9. image capture device as claimed in claim 2, it is characterised in that: the video frequency data frame rate is greater than 5 frames/second.
10. image capture device as claimed in claim 2, it is characterised in that: the two adjacent images in described multiple images Acquisition position at least meet following condition:
H* (1-cosb)=L*sin2b;
A=m*b;
0<m<0.8;
Wherein L is distance of the image collecting device to object, and H is object actual size in acquired image, and a is adjacent Two location drawing picture acquisition device optical axis included angles, m is coefficient.
11. image capture device as claimed in claim 2, it is characterised in that: adjacent three positions in described multiple images Meet three images acquired on corresponding position and at least there is the part for indicating object the same area.
12. a kind of image capture device, it is characterised in that: including
Video data provides device, for providing the video data including object;
Video process apparatus, for handling the video data, to obtain multiple images from the video data;
Processing unit, for obtaining the 3D information of object according at least three in multiple images.
13. a kind of 3D information comparison device, it is characterised in that: set including Image Acquisition described in claim 1-12 any one It is standby.
14. a kind of mating object generating means, it is characterised in that: set using Image Acquisition described in claim 1-12 any one The 3D information at least one region of the standby object obtained generates the mating object matched with object corresponding region.
CN201821449456.6U 2018-09-05 2018-09-05 Image capture device, 3D information comparison device, mating object generating means Active CN208653473U (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201821449456.6U CN208653473U (en) 2018-09-05 2018-09-05 Image capture device, 3D information comparison device, mating object generating means

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201821449456.6U CN208653473U (en) 2018-09-05 2018-09-05 Image capture device, 3D information comparison device, mating object generating means

Publications (1)

Publication Number Publication Date
CN208653473U true CN208653473U (en) 2019-03-26

Family

ID=65792661

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201821449456.6U Active CN208653473U (en) 2018-09-05 2018-09-05 Image capture device, 3D information comparison device, mating object generating means

Country Status (1)

Country Link
CN (1) CN208653473U (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111649690A (en) * 2019-12-12 2020-09-11 天目爱视(北京)科技有限公司 Handheld 3D information acquisition equipment and method
CN113066132A (en) * 2020-03-16 2021-07-02 天目爱视(北京)科技有限公司 3D modeling calibration method based on multi-device acquisition
CN113327291A (en) * 2020-03-16 2021-08-31 天目爱视(北京)科技有限公司 Calibration method for 3D modeling of remote target based on continuous shooting

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111649690A (en) * 2019-12-12 2020-09-11 天目爱视(北京)科技有限公司 Handheld 3D information acquisition equipment and method
CN113066132A (en) * 2020-03-16 2021-07-02 天目爱视(北京)科技有限公司 3D modeling calibration method based on multi-device acquisition
CN113327291A (en) * 2020-03-16 2021-08-31 天目爱视(北京)科技有限公司 Calibration method for 3D modeling of remote target based on continuous shooting
CN113327291B (en) * 2020-03-16 2024-03-22 天目爱视(北京)科技有限公司 Calibration method for 3D modeling of remote target object based on continuous shooting

Similar Documents

Publication Publication Date Title
CN109146949B (en) A kind of 3D measurement and information acquisition device based on video data
CN110580732B (en) 3D information acquisition device
CN111060023B (en) High-precision 3D information acquisition equipment and method
CN109269405B (en) A kind of quick 3D measurement and comparison method
CN109146961B (en) 3D measures and acquisition device based on virtual matrix
CN109141240B (en) A kind of measurement of adaptive 3 D and information acquisition device
CN109285109B (en) A kind of multizone 3D measurement and information acquisition device
CN109218702A (en) A kind of camera rotation type 3D measurement and information acquisition device
CN208653473U (en) Image capture device, 3D information comparison device, mating object generating means
CN208653401U (en) Adapting to image acquires equipment, 3D information comparison device, mating object generating means
CN109443199B (en) 3D information measuring system based on intelligent light source
CN109394168B (en) A kind of iris information measuring system based on light control
CN209279885U (en) Image capture device, 3D information comparison and mating object generating means
CN211178345U (en) Three-dimensional acquisition equipment
CN208795174U (en) Camera rotation type image capture device, comparison device, mating object generating means
CN113066132B (en) 3D modeling calibration method based on multi-equipment acquisition
CN109084679B (en) A kind of 3D measurement and acquisition device based on spatial light modulator
CN111161399B (en) Data processing method and assembly for generating three-dimensional model based on two-dimensional image
CN109394170B (en) A kind of iris information measuring system of no-reflection
CN209103318U (en) A kind of iris shape measurement system based on illumination
CN208795167U (en) Illumination system for 3D information acquisition system
CN213072921U (en) Multi-region image acquisition equipment, 3D information comparison and matching object generation device
CN209279884U (en) Image capture device, 3D information comparison device and mating object generating means
CN113111788A (en) Iris 3D information acquisition equipment with adjusting device
CN109191396A (en) Facial image processing method and apparatus, electronic equipment, computer readable storage medium

Legal Events

Date Code Title Description
GR01 Patent grant
GR01 Patent grant