CN208013783U - The co-located system of the Mobile Robot Formation of view-based access control model - Google Patents

The co-located system of the Mobile Robot Formation of view-based access control model Download PDF

Info

Publication number
CN208013783U
CN208013783U CN201820457284.0U CN201820457284U CN208013783U CN 208013783 U CN208013783 U CN 208013783U CN 201820457284 U CN201820457284 U CN 201820457284U CN 208013783 U CN208013783 U CN 208013783U
Authority
CN
China
Prior art keywords
robot
mobile robot
mobile
formation
access control
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201820457284.0U
Other languages
Chinese (zh)
Inventor
许韶麟
卢洁莹
苏为洲
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201820457284.0U priority Critical patent/CN208013783U/en
Application granted granted Critical
Publication of CN208013783U publication Critical patent/CN208013783U/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
  • Manipulator (AREA)

Abstract

The utility model discloses a kind of co-located systems of the Mobile Robot Formation of view-based access control model, 5cm grades under indoor environment of positioning accuracy is realized when highest movement speed reaches 1m/s, including a base station, three artificial landmarks, four mobile robots and a noncooperative target robot, wherein, base station is used to send control instruction to robot team formation and receives state and location information that robot returns;Artificial landmark is made of the LED of particular color, for establishing cartesian coordinate system;Noncooperative target robot is equipped with the LED of particular color, it is identified for mobile robot;Mobile robot is divided by function autonomous positioning robot and target following robot, wherein, each mobile robot is respectively with there are two cameras, Front camera is for being mutually distinguishable, lock and measuring between mobile robot, then rearmounted camera carries out co-located accordingly for identification, locking and measurement to artificial landmark or noncooperative target.

Description

The co-located system of the Mobile Robot Formation of view-based access control model
Technical field
The utility model is related to the field of locating technology of mobile robot, and in particular to a kind of mobile machine of view-based access control model The co-located system that people forms into columns.
Background technology
Since the twentieth century second half, the research of mobile robot has just obtained quick development, and in national defence, life It is widely used in the fields such as production, industry, service, personal consumption, importance and development prospect are needless to say.Its In, autonomous positioning is a most basic requirement of mobile robot, and such as vision guided navigation, path planning, target are surrounded and seize Etc. the basis of more advanced function, while it being also the necessary condition that mobile robot possesses " intelligence ".But robot is certainly at this stage The research of master positioning often concentrates on individual machine people.And under certain application backgrounds, individual machine people is in sensing range, meter Calculation ability, processing capacity, anti-interference ability etc. have certain defect, autonomous positioning precision to tend not to very high.With it is single Robot compares, and possesses more perception units by the robot team formation that multirobot forms, can improve sensing range;This Outside, the increase of information content can make robot team formation obtain higher positioning accuracy;Changeable topological structure makes robot compile Team can execute more complicated algorithm, while Distributed Parallel Computing can make system possess stronger computing capability;In addition, compiling Robot in team can provide redundancy, to possess outstanding anti-interference ability by mutually coordinated for the completion of task.
Utility model content
The purpose of this utility model is to overcome the deficiencies of individual machine people autonomous positioning at this stage, are subsequent movement machine Device man-based development lays the first stone, it is proposed that a kind of co-located system of the Mobile Robot Formation of view-based access control model, the system Mobile robot has merged a variety of heat transfer agents and communication module based on visual information, reaches 1m/ in highest movement speed In the case of s, 5cm grades under environment indoors of positioning accuracy can be realized.
The purpose of this utility model can be reached by adopting the following technical scheme that:
A kind of co-located system of the Mobile Robot Formation of view-based access control model, the co-located system include:One A base station, three artificial landmarks, four mobile robots and a noncooperative target robot, above-mentioned base station, artificial landmark, shifting Mobile robot and noncooperative target robot are equipped with wireless module, for carrying out mutual wireless communication;
Wherein, the base station be used for robot team formation send control instruction and receive robot return state and Location information;The artificial landmark is made of the LED of particular color, for establishing cartesian coordinate system;The non-cooperation Target robot is equipped with the LED of particular color, it is identified for mobile robot;Mobile robot described in four is pressed Function divides as follows:Two mobile robots are as autonomous positioning robot, and two mobile robots are as target following machine People, wherein the mobile robot is respectively matched there are two camera, the wherein Front camera shifting adjacent with locking for identification LED on mobile robot, and relative distance and relative angle with adjacent mobile robot are measured, rearmounted camera is used for Identification and locking artificial landmark or noncooperative target robot, and measure and locking artificial landmark or noncooperative target robot Relative angle.
Further, each mobile robot is made of driving layer, sensing layer, control layer, wherein the drive Dynamic layer is made of wheel, direct current generator, DC motor driver, absolute value encoder, the movement for driving mobile robot; The sensing layer includes preposition Vision Table, postposition Vision Table, laser range sensor, gesture module, for obtaining machine Device people movable information and environmental information,;The control layer includes industrial personal computer, K60 microcontrollers, and wherein industrial personal computer is as main control Device, for obtaining camera image information and realizing co-located function, K60 microcontrollers are used as from controller, for vision Holder is controlled, and so that it is locked and is tracked noncooperative target robot.
Further, the top of each mobile robot is equipped with the LED of particular color, as robot identity Mark.
Further, the preposition Vision Table of the mobile robot is single-degree-of-freedom.
Further, the postposition Vision Table of the autonomous positioning robot is single-degree-of-freedom, the target following The postposition Vision Table of robot is two degrees of freedom.
Further, each mobile robot is carried out wireless communication with base station by way of time-slotting, real The interaction of existing control instruction, posture information, movable information, heat transfer agent.
The utility model has the following advantages compared with the existing technology and effect:
1) the utility model designs and has made four simple in structure, powerful mobile robots, can realize pair Identification, tracking and the locking of specific objective;
2) the utility model designs and makes a set of co-located system for robot team formation, in highest movement speed In the case of reaching 1m/s, robot can realize 5cm grades under environment indoors of positioning accuracy;
3) sophisticated functions such as the opening of this system allows target following, target to surround and seize, map structuring, environment are searched for Follow-up addition.
Description of the drawings
Fig. 1 be the Mobile Robot Formation of view-based access control model disclosed by the utility model co-located system in mobile machine The structure chart of people;
Fig. 2 is the structure chart of the co-located system of the Mobile Robot Formation of view-based access control model disclosed by the utility model;
Fig. 3 is the controlling cycle of the co-located system of the Mobile Robot Formation of view-based access control model disclosed by the utility model With the schematic diagram of communication cycle.
Specific implementation mode
It is new below in conjunction with this practicality to keep the purpose, technical scheme and advantage of the utility model embodiment clearer Attached drawing in type embodiment, the technical scheme in the utility model embodiment is clearly and completely described, it is clear that is retouched The embodiment stated is the utility model a part of the embodiment, instead of all the embodiments.Based on the implementation in the utility model Example, every other embodiment obtained by those of ordinary skill in the art without making creative efforts belong to The range of the utility model protection.
Embodiment
The utility model discloses a kind of co-located system of the Mobile Robot Formation of view-based access control model, which includes One base station, three artificial landmarks (being hereafter together simply referred to as road sign), four mobile robots (can be divided into two by function division Autonomous positioning robot and Liang Ge target followings robot, hereafter together simply referred to as robot) with a noncooperative target machine People (is hereafter together simply referred to as target).Wherein, base station is the computer of an operation application-specific, for being compiled to robot Team sends control instruction and receives state and location information that robot returns;Road sign is mainly made of the LED of particular color, Its main function is to establish cartesian coordinate system;Target is exactly a telecar in simple terms, can freely be moved.Mesh The LED equipped with particular color is put on, facilitates robot that it is identified.
From the point of view of structure, each robot can be divided into driving layer, sensing layer, control layer.Drive layer mainly by wheel, The compositions such as direct current generator, DC motor driver, absolute value encoder, the movement for driving robot;Sensing layer mainly by Two Vision Tables (preposition Vision Table tracking road sign or target, postposition Vision Table track mutually), laser range sensors The compositions such as (being assemblied on preposition Vision Table, moved with camera), gesture module, for obtain robot motion's information with Environmental information;Control layer mainly by industrial personal computer, K60 singlechip groups at.Industrial personal computer is as master controller, for obtaining camera image Information and operation co-located algorithm.K60 microcontrollers, which are used as from controller for controlling Vision Table, makes its lock Determine and tracks target.In addition, the top of each robot is equipped with the LED of particular color, the mark as robot identity.From (autonomous positioning robot is equipped with single for master positioning robot and the postposition Vision Table that differs only in of target following robot architecture The Vision Table of degree of freedom, target following robot are furnished with the Vision Table of two degrees of freedom).
In addition to this, base station, artificial landmark and each robot are provided with wireless module, for carrying out mutual nothing Line communicates.
Include mainly two stages when this system is run:Initial alignment stage and motion positions stage.In initial alignment rank Section sends signal by base station and controls three road signs and lights in turn, at this time the rearmounted camera of each robot to the road sign lighted into Row locking, and obtain opposite yaw angle;In addition, the Front camera of each robot carries out mutually locking, obtain between robot Opposite yaw angle and relative distance.Then non-linear least square problem is built, and is solved using gauss-newton method, is obtained each The initial world coordinates of a robot.The initial alignment stage terminates.
In the motion positions stage, only there are one the road sign fixations as coordinate origin to light.Liang Ge autonomous positionings robot Front camera lock mutually, rearmounted camera lock artificial landmark;The Front camera locking of Liang Ge target followings robot is neighbouring Autonomous positioning robot, rearmounted camera lock noncooperative target.After base station sends setting in motion instruction, four robots open Begin to move.On the basis of initial alignment, robot interacts respective posture information, movable information and biography by time-slotting Feel information, then by Extended Kalman filter (EKF) algorithm to oneself, other robots and noncooperative target determine Position.
It can be seen in fig. 1 that from the angle of hardware module, robot is by industrial personal computer, motion module, preposition Vision Table Module, postposition Vision Table module and wireless communication module composition.Wherein, industrial personal computer is mainly used for obtaining as master controller Take camera image information, robot pose information, operation co-located algorithm and and base station communication;Motion module is mono- by K60 Piece machine, DC motor driver, direct current generator and photoelectric encoder form.K60 microcontrollers receive industrial personal computer by serial ports and send Instruction, control wheel runs with specified speed, is then back to the speed of wheel to industrial personal computer;Similarly, forward and backward to set vision Holder module is made of K60 microcontrollers, stepper motor driver, stepper motor, absolute value encoder and camera.K60 monolithics Machine receives the instruction that industrial personal computer is sent by serial ports, and control Vision Table makes camera track and locks specific objective, then returns The state of holder and camera is returned to industrial personal computer.
Fig. 2 is the system construction drawing of robot team formation co-located.It can be seen that whole system is by base station, road sign, target And four robot F1, F2, F3 and F4 compositions (function-differentiated, F1, F2 are autonomous positioning robots, F3, F4 be target with Track robot).Wherein, other than the modules described in Fig. 1, the LED as identity is also equipped in each robot. In figure, base station runs upper computer software, issues control instruction to four robots by wireless module, while receiving four machines The location information that people returns, shows on software.The preposition Vision Table module of four robots can be identified and be locked LED in adjacent robot, and measure relative distance and relative angle with adjacent machines people;Postposition Vision Table module Road sign or target can be identified and be locked, and measures the relative angle with road sign or target.It is described in detail by taking F1 as an example, before F1 The LED of the camera identification F2 of Vision Table module is set, stepper motor drives preposition Vision Table rotation so that camera is tracked and locked Determine F2.Then, the laser range sensor in preposition Vision Table module and absolute value encoder measure opposite between F2 Distance and relative angle.Similarly, the camera of postposition Vision Table module identifies that road sign, stepper motor drive postposition Vision Table Rotation so that camera tracks and locks road sign.Then, absolute value encoder in postposition Vision Table module measure with road sign it Between relative angle.Similarly, the preposition Vision Table module locking of F2 is locked and is surveyed with F1, postposition Vision Table module is measured Measure road sign;The preposition Vision Table module locking of F3 locks with F1, the preposition Vision Table module of F4 is measured and measures F2, they Postposition Vision Table module then lock with measure target.In addition to this, robot can also obtain movable information from motion module, Posture information is obtained from gesture module.In summary four robots and mesh may be implemented in the information such as observation, movement, posture Target co-located.
Fig. 3 is the schematic diagram of system controlling cycle and communication cycle.The control of robot F1, F2, F3 and F4 in this system Period processed is all 36ms.In controlling cycle, robot needs to complete acquisition image, processing image, filtering positioning, path rule It draws, the work such as motion control.In addition, since robot and base station are all the wireless communications using broadcast type, if not to system Communication cycle carry out time planning, it will collision and the loss for causing the communication information, cause system perturbations with failure.Therefore, This system is managed communication cycle by the way of dividing timeslice.Specifically, in the starting of each communication cycle, base It stands first broadcast transmission control instruction, the call duration time for distributing to base station is 24ms;After the information for receiving base station broadcast, F1 is at once The information such as state, pose and the observation of broadcast transmission itself, F2, F3 and F4 are sent out again after then waiting for 30ms, 60ms, 90ms respectively It send.This means that the timeslice for distributing to each robot is 30ms.And in the respective timeslice of robot, only should Robot is in broadcast transmission data, information collision caused by being wirelessly communicated so as to avoid broadcast type and loss.In addition to this, machine 4 controlling cycles of device people are just equal to 1 complete communication cycle, realize being aligned for controlling cycle and communication cycle.
Above-described embodiment is the preferable embodiment of the utility model, but the embodiment of the utility model is not by above-mentioned The limitation of embodiment, under other any Spirit Essences and principle without departing from the utility model made by change, modify, replace In generation, simplifies combination, should be equivalent substitute mode, is included within the scope of protection of the utility model.

Claims (6)

1. a kind of co-located system of the Mobile Robot Formation of view-based access control model, which is characterized in that the co-located system System includes:One base station, three artificial landmarks, four mobile robots and a noncooperative target robot, above-mentioned base station, people Work road sign, mobile robot and noncooperative target robot are equipped with wireless module, for carrying out mutual wireless communication;
Wherein, the base station is used to send control instruction to robot team formation and receives state and positioning that robot returns Information;The artificial landmark is made of the LED of particular color, for establishing cartesian coordinate system;The noncooperative target Robot is equipped with the LED of particular color, it is identified for mobile robot;Mobile robot described in four presses function It divides as follows:Two mobile robots as autonomous positioning robot, two mobile robots as target following robot, In, the mobile robot is respectively matched there are two camera, wherein the Front camera moving machine adjacent with locking for identification LED on device people, and measure with the relative distance and relative angle of adjacent mobile robot, rearmounted camera for identification with lock Determine artificial landmark or noncooperative target robot, and measures the relative angle with locking artificial landmark or noncooperative target robot Degree.
2. the co-located system of the Mobile Robot Formation of view-based access control model according to claim 1, which is characterized in that every A mobile robot is made of driving layer, sensing layer, control layer, wherein the driving layer is by wheel, direct current Machine, DC motor driver, absolute value encoder composition, the movement for driving mobile robot;The sensing layer includes Preposition Vision Table, postposition Vision Table, laser range sensor, gesture module, for obtaining robot motion's information and ring Border information, the control layer include industrial personal computer, K60 microcontrollers, and wherein industrial personal computer is as master controller, for obtaining camera figure As information and realize co-located function, K60 microcontrollers, which are used as from controller for controlling Vision Table, makes it It locks and tracks noncooperative target robot.
3. the co-located system of the Mobile Robot Formation of view-based access control model according to claim 1, which is characterized in that every The top of a mobile robot is equipped with the LED of particular color, the mark as robot identity.
4. the co-located system of the Mobile Robot Formation of view-based access control model according to claim 2, which is characterized in that institute The preposition Vision Table for the mobile robot stated is single-degree-of-freedom.
5. the co-located system of the Mobile Robot Formation of view-based access control model according to claim 2, which is characterized in that institute The postposition Vision Table of the autonomous positioning robot stated is single-degree-of-freedom, the postposition Vision Table of the target following robot For two degrees of freedom.
6. the co-located system of the Mobile Robot Formation of view-based access control model according to claim 1, which is characterized in that every A mobile robot is carried out wireless communication with base station by way of time-slotting, realization control instruction, posture information, The interaction of movable information, heat transfer agent.
CN201820457284.0U 2018-04-03 2018-04-03 The co-located system of the Mobile Robot Formation of view-based access control model Expired - Fee Related CN208013783U (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201820457284.0U CN208013783U (en) 2018-04-03 2018-04-03 The co-located system of the Mobile Robot Formation of view-based access control model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201820457284.0U CN208013783U (en) 2018-04-03 2018-04-03 The co-located system of the Mobile Robot Formation of view-based access control model

Publications (1)

Publication Number Publication Date
CN208013783U true CN208013783U (en) 2018-10-26

Family

ID=63891824

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201820457284.0U Expired - Fee Related CN208013783U (en) 2018-04-03 2018-04-03 The co-located system of the Mobile Robot Formation of view-based access control model

Country Status (1)

Country Link
CN (1) CN208013783U (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109839111A (en) * 2019-01-10 2019-06-04 王昕� A kind of indoor multi-robot formation system of view-based access control model positioning
CN110007671A (en) * 2019-03-12 2019-07-12 南方科技大学 A kind of robot cluster co-located system and method
CN110147097A (en) * 2019-04-28 2019-08-20 深兰科技(上海)有限公司 A kind of location determining method of dispensing device, device, equipment and medium
CN110162094A (en) * 2019-06-13 2019-08-23 中国人民解放军军事科学院国防科技创新研究院 A kind of close/intra control method of view-based access control model metrical information
CN111077889A (en) * 2019-12-23 2020-04-28 西北工业大学 Multi-mobile-robot formation cooperative positioning method for workshop tray transportation
CN111185904A (en) * 2020-01-09 2020-05-22 上海交通大学 Collaborative robot platform and control system thereof
CN112731923A (en) * 2020-12-17 2021-04-30 武汉万集信息技术有限公司 Cluster robot cooperative positioning system and method
CN112985259A (en) * 2021-01-25 2021-06-18 中国人民解放军军事科学院国防科技创新研究院 Target positioning method and system based on multi-view vision

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109839111A (en) * 2019-01-10 2019-06-04 王昕� A kind of indoor multi-robot formation system of view-based access control model positioning
CN110007671A (en) * 2019-03-12 2019-07-12 南方科技大学 A kind of robot cluster co-located system and method
CN110147097A (en) * 2019-04-28 2019-08-20 深兰科技(上海)有限公司 A kind of location determining method of dispensing device, device, equipment and medium
CN110162094A (en) * 2019-06-13 2019-08-23 中国人民解放军军事科学院国防科技创新研究院 A kind of close/intra control method of view-based access control model metrical information
CN110162094B (en) * 2019-06-13 2022-02-22 中国人民解放军军事科学院国防科技创新研究院 Intensive formation control method based on visual measurement information
CN111077889A (en) * 2019-12-23 2020-04-28 西北工业大学 Multi-mobile-robot formation cooperative positioning method for workshop tray transportation
CN111077889B (en) * 2019-12-23 2023-04-14 西北工业大学 Multi-mobile-robot formation cooperative positioning method for workshop tray transportation
CN111185904A (en) * 2020-01-09 2020-05-22 上海交通大学 Collaborative robot platform and control system thereof
CN112731923A (en) * 2020-12-17 2021-04-30 武汉万集信息技术有限公司 Cluster robot cooperative positioning system and method
CN112731923B (en) * 2020-12-17 2023-10-03 武汉万集光电技术有限公司 Cluster robot co-positioning system and method
CN112985259A (en) * 2021-01-25 2021-06-18 中国人民解放军军事科学院国防科技创新研究院 Target positioning method and system based on multi-view vision
CN112985259B (en) * 2021-01-25 2023-05-16 中国人民解放军军事科学院国防科技创新研究院 Target positioning method and system based on multi-view vision

Similar Documents

Publication Publication Date Title
CN208013783U (en) The co-located system of the Mobile Robot Formation of view-based access control model
US10800036B1 (en) Tooltip stabilization
CN103353758B (en) A kind of Indoor Robot navigation method
CN105425791B (en) A kind of the group robot control system and method for view-based access control model positioning
CN107544515A (en) Multirobot based on Cloud Server builds figure navigation system and builds figure air navigation aid
CN104834309B (en) The optimal touring control method of single mobile robot based on target following control strategy
CN102419178B (en) Mobile robot positioning system and method based on infrared road sign
CN109062204A (en) It is a kind of based on follow pilotage people form into columns multiple mobile robot's control system
CN103659817B (en) A kind of rescuing robot system and realize the method for this rescue system
US11230016B1 (en) Multi-resolution localization system
CN100461058C (en) Automatic positioning method for intelligent robot under complex environment
Yuan et al. RGB-D sensor-based visual SLAM for localization and navigation of indoor mobile robot
CN105116886B (en) A kind of method of robot autonomous walking
CN103064416A (en) Indoor and outdoor autonomous navigation system for inspection robot
CN109144068A (en) The automatically controlled mode and control device of three-dimensional shift-forward type navigation switching AGV fork truck
CN106383517A (en) Control system used for autonomous mobile robot platform and control method and device thereof
CN107562054A (en) The independent navigation robot of view-based access control model, RFID, IMU and odometer
CN107621826A (en) Intelligent movable tracks car
CN207473406U (en) A kind of intelligent carriage tracking system with Quick Response Code site location information
CN108255173A (en) Robot follows barrier-avoiding method and device
CN111077889A (en) Multi-mobile-robot formation cooperative positioning method for workshop tray transportation
CN108107884A (en) Robot follows the data processing method and its intelligent apparatus of navigation
CN110530371A (en) A kind of indoor map matching process based on deeply study
CN111113415B (en) Robot positioning method based on two-dimensional code road sign, camera and gyroscope
Cheng et al. Person-following for telepresence robots using web cameras

Legal Events

Date Code Title Description
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20181026

Termination date: 20200403

CF01 Termination of patent right due to non-payment of annual fee