CN102338621A - Method for detecting height of obstacle for indoor visual navigation - Google Patents
Method for detecting height of obstacle for indoor visual navigation Download PDFInfo
- Publication number
- CN102338621A CN102338621A CN2011101053892A CN201110105389A CN102338621A CN 102338621 A CN102338621 A CN 102338621A CN 2011101053892 A CN2011101053892 A CN 2011101053892A CN 201110105389 A CN201110105389 A CN 201110105389A CN 102338621 A CN102338621 A CN 102338621A
- Authority
- CN
- China
- Prior art keywords
- ground
- camera
- obstacle
- point
- indoor
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Landscapes
- Image Processing (AREA)
- Navigation (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
Abstract
The invention discloses a method for detecting the height of a ground obstacle in indoor environment, belonging to the technical field of visual navigation. The method can help an indoor visual navigation system to judge the obstacle area and the ground area in the indoor environment correctly. The invention is realized by the following technical scheme that a camera and a laser transmitting device are arranged on an indoor ceiling, wherein the laser transmitting device can rotate randomly around the axis. The optical center of the camera and the rotating axis of the laser transmitting device are positioned in the same horizontal surface. The laser transmitting device projects a light spot to the ground, and the light spot can reach the ground if no obstacle exists; and the light spot is projected on the obstacle if the obstacle exists. According to the method, an indoor top image is shot by utilizing the camera, and the height of the obstacle is calculated by utilizing the offset of the imaging positions of the light spots projected to the ground and the obstacle.
Description
Technical field
The present invention belongs to the vision guided navigation technical field, relates to a kind of method that in indoor environment, detects the ground obstacle height.
Background technology
Vision guided navigation is meant the information of utilizing imageing sensor to obtain environment, realizes robot navigation's technology through the analysis environments image.Vision guided navigation obtains the mode of information with human identical, therefore is considered to have one of airmanship of development potentiality most.
In indoor environment, can be fixed on imageing sensor on the ceiling to obtain indoor overhead view image, realize the navigation of robot through analyzing overhead view image then.In this navigate mode, need cut apart the ground region of overhead view image, through the activity restriction of robot in ground region, realize hiding automatically to barrier.Existing ground region dividing method generally adopts visual signatures such as the color, texture on ground to extract ground region, has following two shortcomings:
(1), for example is placed on white ground white chair etc., then correctly the ground region of split image and barrier region if the visual signature ground region of barrier is identical;
(2) if the visual signature of ground region is inconsistent, for example there is a beach ink marks etc. on ground, can cause vision algorithm to be barrier region to the ground region erroneous judgement.
Summary of the invention
Technical matters to be solved by this invention is the above-mentioned deficiency to prior art, proposes a kind of indoor Visual Navigation and uses the obstacle height detection method.
The present invention adopts following technical scheme, comprises the following steps:
Camera and laser beam emitting device are installed on the indoor ceiling.The rotatable axle center C of laser beam emitting device rotates arbitrarily, and horizontal direction rotation corner is expressed as θ, and vertically the aspect corner is φ.The photocentre O of axle center C and camera is positioned at same surface level, and the distance that promptly arrives ground is identical.The distance table that the O point is ordered to C is shown
O is to O
fThe distance table of point is shown
With
Mode through actual measurement obtains.The focal length of camera is f
c
Known level rotational angle theta and vertical plane corner φ, laser beam emitting device is projected spot earthward.If there is not barrier, then hot spot projects the R point on ground.The coordinate that R is ordered calculates according to
and θ, φ.The R point at camera as the i that is imaged as in the plane
R, utilize aperture camera model i
RCoordinate by the focal distance f of camera
c,
And the coordinate Calculation of R obtains.If there is barrier in ground, then hot spot can not arrive ground, but projects the P point of barrier, and the P point is to the distance on ground, i.e. the height h of barrier.The P point is at the imaging i on picture plane
PUtilize camera to take the overhead view image of indoor environment, i
PCoordinate obtain through the position of analyzing hot spot in the overhead view image.Picture i
PWith picture i
RBetween distance table be shown
The height h of barrier by
f
cWith
Calculate.
Because in actual application, there is the erroneous judgement phenomenon in original method to barrier region in the indoor environment and ground region, promptly is ground region to the barrier region erroneous judgement that should dodge, is barrier region to the ground region erroneous judgement that can pass through.In order to address this problem, the present invention's proposition utilizes the mode of active vision to obtain the height of indoor diverse location, and whether utilize elevation information to judge should the zone be barrier.This method can be united use with original method, can improve the accuracy rate of navigational system cognitive disorders thing.
Description of drawings
Fig. 1 illustrates the installation site of camera and laser beam emitting device;
Among the figure: 1. camera; 2. laser beam emitting device
Fig. 2 illustrates the geometric relationship of dyscalculia object height degree.
Among the figure: 1. as the plane; 2. laser beam emitting device; 3. barrier
Embodiment
1. the installation of camera and laser beam emitting device
Be installed in camera and laser beam emitting device on the ceiling, laser beam emitting device can rotate in two degree of freedom around the axle center.O representes the photocentre of camera among Fig. 1, and the optical axis of camera is perpendicular to the ground, and with the intersection point on ground be O
fC representes the axle center of laser beam emitting device rotation, and the axis of along continuous straight runs rotation is perpendicular to the ground, and with the intersection point on ground be C
fLaser beam emitting device horizontal direction rotation corner is expressed as θ, and vertically the aspect corner is φ.Photocentre O and axle center C are positioned at same surface level, and promptly 2 distances to ground equate.
The distance table that the O point is ordered to C is shown
O is to O
fThe distance table of point is shown
With
Mode through actual measurement obtains.The focal length of camera is f
c
Under the known condition of θ and φ, hot spot coordinate on the ground is R (x
R, y
R, 0), x wherein
R, and y
R, calculate by formula (1) and (2)
Utilize the pinhole imaging system model, can obtain R picture i in the picture plane
RCoordinate
2. obtain the volume coordinate of hot spot imaging
Take indoor overhead view image, when having barrier on the ground, hot spot can not arrive the R point on ground, but projects the P point of barrier.The P point is at the i that is imaged as as the plane
PUtilize image process method to obtain i
PCoordinate.
(1) under the known situation of corner φ and θ, it is (R that laser beam emitting device throws color value earthward
l, G
l, B
l) monochromatic hot spot, R
l, G
l, B
lThe red component of representing color value respectively, green component and blue component.
(2) be W as the rectangular area that is used in the plane form images, because the graphical representation that the z coordinate is
therefore omits z coordinate handle of in as the plane, being had a few does
I=[I
r(x,y),I
g(x,y),I
b(x,y)|(x,y)∈W]
I wherein
r(x, y), I
g(x, y), I
b(x y) is respectively image at (x, y) red component, green component and the blue component of coordinate place color value.
(3) select threshold value T, utilize following formula that light spot image is cut apart:
(4) according to the central point i of computes hot spot
PCoordinate (x
P, y
P)
3. obstacle height is calculated
In Fig. 2; The picture plane is parallel to the ground; Simultaneously optical axis is O ' with intersection point as the plane, and the distance table between O ' and O ' is shown
and is the focal length of camera.
Do vertical line earthward by the P point; Give the B point with ground; Then
representes the distance between P point and the B point, and
h is the height of barrier.The extended line and the ground of O point and P point line meet at P ' point.i
RAnd i
PBe respectively R and the P picture in the picture plane
Utilize the aperture camera model to obtain
Similar according to triangle RPP ' with CPO, and BPP ' and O
fOP ' is similar, is concerned as follows
Aggregative formula (7), (8) and (9) obtain,
Claims (1)
1. a method that in indoor environment, detects the ground obstacle height comprises the following steps:
(1) be installed in camera and laser beam emitting device on the ceiling, the axis of rotation C of laser beam emitting device and the photocentre O of camera are at same surface level; The distance that the O point is ordered to C does
The optical axis of camera is perpendicular to the ground, and meets at O with ground
fThe point; O is to O
fThe distance of point does
Be the distance of O point to ground; The focal length of camera is f
c
(2) anglec of rotation of fixed laser emitter, launch spot earthward; When not having barrier on the ground, hot spot projects on the ground, at the be imaged as i of camera as the plane
RWhen on the ground barrier being arranged, hot spot projects on the barrier, at the be imaged as i of camera as the plane
Pi
RAnd i
PBetween have skew, be expressed as
Utilize the height h of the geometry imaging relations dyscalculia thing of aperture camera model:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2011101053892A CN102338621B (en) | 2011-04-27 | 2011-04-27 | Method for detecting height of obstacle for indoor visual navigation |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2011101053892A CN102338621B (en) | 2011-04-27 | 2011-04-27 | Method for detecting height of obstacle for indoor visual navigation |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102338621A true CN102338621A (en) | 2012-02-01 |
CN102338621B CN102338621B (en) | 2013-11-20 |
Family
ID=45514473
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN2011101053892A Expired - Fee Related CN102338621B (en) | 2011-04-27 | 2011-04-27 | Method for detecting height of obstacle for indoor visual navigation |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102338621B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106500699A (en) * | 2016-05-25 | 2017-03-15 | 上海铸天智能科技有限公司 | A kind of position and orientation estimation method suitable for Autonomous landing in unmanned plane room |
CN107728633A (en) * | 2017-10-23 | 2018-02-23 | 广州极飞科技有限公司 | Obtain object positional information method and device, mobile device and its control method |
CN109945790A (en) * | 2017-12-21 | 2019-06-28 | 中国科学院长春光学精密机械与物理研究所 | A kind of optical means for establishing characteristic point for large-scale upright object |
CN111506074A (en) * | 2020-05-08 | 2020-08-07 | 佳木斯大学 | Machine control method of crop tedding dust collection device |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060041333A1 (en) * | 2004-05-17 | 2006-02-23 | Takashi Anezaki | Robot |
CN101619985A (en) * | 2009-08-06 | 2010-01-06 | 上海交通大学 | Service robot autonomous navigation method based on deformable topological map |
CN101975951A (en) * | 2010-06-09 | 2011-02-16 | 北京理工大学 | Field environment barrier detection method fusing distance and image information |
-
2011
- 2011-04-27 CN CN2011101053892A patent/CN102338621B/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060041333A1 (en) * | 2004-05-17 | 2006-02-23 | Takashi Anezaki | Robot |
CN101619985A (en) * | 2009-08-06 | 2010-01-06 | 上海交通大学 | Service robot autonomous navigation method based on deformable topological map |
CN101975951A (en) * | 2010-06-09 | 2011-02-16 | 北京理工大学 | Field environment barrier detection method fusing distance and image information |
Non-Patent Citations (1)
Title |
---|
张博等: "室内移动机器人新型视觉导航***设计", 《工业控制计算机》 * |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106500699A (en) * | 2016-05-25 | 2017-03-15 | 上海铸天智能科技有限公司 | A kind of position and orientation estimation method suitable for Autonomous landing in unmanned plane room |
CN106500699B (en) * | 2016-05-25 | 2019-06-18 | 上海铸天智能科技有限公司 | A kind of position and orientation estimation method suitable for Autonomous landing in unmanned plane room |
CN107728633A (en) * | 2017-10-23 | 2018-02-23 | 广州极飞科技有限公司 | Obtain object positional information method and device, mobile device and its control method |
CN109945790A (en) * | 2017-12-21 | 2019-06-28 | 中国科学院长春光学精密机械与物理研究所 | A kind of optical means for establishing characteristic point for large-scale upright object |
CN109945790B (en) * | 2017-12-21 | 2020-09-01 | 中国科学院长春光学精密机械与物理研究所 | Optical method for establishing characteristic points for large vertical object |
CN111506074A (en) * | 2020-05-08 | 2020-08-07 | 佳木斯大学 | Machine control method of crop tedding dust collection device |
CN111506074B (en) * | 2020-05-08 | 2022-08-26 | 佳木斯大学 | Machine control method of crop tedding dust collection device |
Also Published As
Publication number | Publication date |
---|---|
CN102338621B (en) | 2013-11-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11865708B2 (en) | Domestic robotic system | |
CN102650886B (en) | Vision system based on active panoramic vision sensor for robot | |
CN105300375B (en) | A kind of robot indoor positioning and air navigation aid based on single vision | |
CN104299240B (en) | Camera calibration method and system for lane shift early warning | |
CN103076802B (en) | Robot virtual boundary is set up and recognition methods and system | |
KR101703177B1 (en) | Apparatus and method for recognizing position of vehicle | |
US7684894B2 (en) | Autonomously moving robot | |
CN102419178B (en) | Mobile robot positioning system and method based on infrared road sign | |
CN108663681A (en) | Mobile Robotics Navigation method based on binocular camera Yu two-dimensional laser radar | |
CN102338621B (en) | Method for detecting height of obstacle for indoor visual navigation | |
CN106909145A (en) | Unmanned hydrographical survey ship barrier real-time perception obstacle avoidance system and method | |
CN106855411A (en) | A kind of robot and its method that map is built with depth camera and obstacle avoidance system | |
CN104061865B (en) | A kind of navigating ship altitude measurement system and measuring method thereof | |
CN102175991B (en) | Target positioning method based on maximum positioning likelihood sensor configuration | |
CN106569225A (en) | Range-finding sensor based real-time obstacle avoidance method of driveless car | |
CN113289290B (en) | Fire-fighting robot flame automatic aiming method, device and system | |
CN105389543A (en) | Mobile robot obstacle avoidance device based on all-dimensional binocular vision depth information fusion | |
KR102031348B1 (en) | Autonomous Working System, Method and Computer Readable Recording Medium | |
US10348961B2 (en) | Camera modeling system | |
WO2018129648A1 (en) | Robot and method thereof for establishing map by using deep camera and obstacle avoidance system | |
CN104036518A (en) | Camera calibration method based on vector method and three collinear points | |
Yuan et al. | Real-time security monitoring around a video surveillance vehicle with a pair of two-camera omni-imaging devices | |
Eguchi et al. | Development of the autonomous mobile robot for target-searching in urban areas in the Tsukuba Challenge 2013 | |
CN108088375A (en) | A kind of object relative position detection light beam method of reseptance and device | |
CN106123864B (en) | Image distance measuring method based on image-forming principle and Data Regression Model |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
C17 | Cessation of patent right | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20131120 Termination date: 20140427 |