EP3147743B1 - Indoor monitoring system and method for structure - Google Patents

Indoor monitoring system and method for structure Download PDF

Info

Publication number
EP3147743B1
EP3147743B1 EP15814504.5A EP15814504A EP3147743B1 EP 3147743 B1 EP3147743 B1 EP 3147743B1 EP 15814504 A EP15814504 A EP 15814504A EP 3147743 B1 EP3147743 B1 EP 3147743B1
Authority
EP
European Patent Office
Prior art keywords
distance
floating machine
unmanned floating
position information
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
EP15814504.5A
Other languages
German (de)
French (fr)
Other versions
EP3147743A4 (en
EP3147743A1 (en
Inventor
Takayuki Kono
Yoichiro Tsumura
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mitsubishi Heavy Industries Ltd
Original Assignee
Mitsubishi Heavy Industries Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mitsubishi Heavy Industries Ltd filed Critical Mitsubishi Heavy Industries Ltd
Publication of EP3147743A1 publication Critical patent/EP3147743A1/en
Publication of EP3147743A4 publication Critical patent/EP3147743A4/en
Application granted granted Critical
Publication of EP3147743B1 publication Critical patent/EP3147743B1/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/0011Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots associated with a remote control arrangement
    • G05D1/0038Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots associated with a remote control arrangement by providing the operator with simple or augmented images from one or more cameras located onboard the vehicle, e.g. tele-operation
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64DEQUIPMENT FOR FITTING IN OR TO AIRCRAFT; FLIGHT SUITS; PARACHUTES; ARRANGEMENT OR MOUNTING OF POWER PLANTS OR PROPULSION TRANSMISSIONS IN AIRCRAFT
    • B64D47/00Equipment not otherwise provided for
    • B64D47/08Arrangements of cameras
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B21/00Measuring arrangements or details thereof, where the measuring technique is not covered by the other groups of this subclass, unspecified or not relevant
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C15/00Surveying instruments or accessories not provided for in groups G01C1/00 - G01C13/00
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/20Instruments for performing navigational calculations
    • G01C21/206Instruments for performing navigational calculations specially adapted for indoor navigation
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C23/00Combined instruments indicating more than one navigational value, e.g. for aircraft; Combined measuring devices for measuring two or more variables of movement, e.g. distance, speed or acceleration
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C3/00Measuring distances in line of sight; Optical rangefinders
    • G01C3/02Details
    • G01C3/06Use of electric means to obtain final indication
    • G01C3/08Use of electric radiation detectors
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/84Systems specially adapted for particular applications
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/0094Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots involving pointing a payload, e.g. camera, weapon, sensor, towards a fixed or moving target
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64CAEROPLANES; HELICOPTERS
    • B64C39/00Aircraft not otherwise provided for
    • B64C39/02Aircraft not otherwise provided for characterised by special use
    • B64C39/024Aircraft not otherwise provided for characterised by special use of the remote controlled vehicle type, i.e. RPV
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U10/00Type of UAV
    • B64U10/10Rotorcrafts
    • B64U10/13Flying platforms
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U10/00Type of UAV
    • B64U10/10Rotorcrafts
    • B64U10/13Flying platforms
    • B64U10/14Flying platforms with four distinct rotor axes, e.g. quadcopters
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U2101/00UAVs specially adapted for particular uses or applications
    • B64U2101/30UAVs specially adapted for particular uses or applications for imaging, photography or videography
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U2201/00UAVs characterised by their flight controls
    • B64U2201/20Remote controls
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U30/00Means for producing lift; Empennages; Arrangements thereof
    • B64U30/20Rotors; Rotor supports

Definitions

  • the present invention relates to an indoor monitoring system and method for a structure.
  • a boiler furnace used at a thermal power plant needs to be opened during construction and periodically after starting operation so that a worker enters the inside to conduct maintenance inspection.
  • this maintenance inspection it is necessary to define a portion to be inspected, but it is difficult to accurately grasp the portion to be inspected visually because the capacity of the boiler furnace is large.
  • a height position and a lateral position of the portion to be inspected have been conventionally measured and marked using a measuring tape or the like to grasp where the worker is or a maintenance inspection position, but such a method requires not only erection of scaffolding for the worker and installation of a gondola but also a lot of efforts, cost, and inspection periods.
  • Patent Literature 1 a technique has been conventionally proposed to clean up the inside of a structure such as a stack using an unmanned inspection apparatus (Patent Literature 1).
  • Patent Literature 1 a technique has been conventionally proposed to clean up the inside of a structure such as a stack using an unmanned inspection apparatus (Patent Literature 1).
  • Patent Literature 1 a technique has been conventionally proposed to clean up the inside of a structure such as a stack using an unmanned inspection apparatus.
  • Patent Literature 1 an unmanned inspection apparatus
  • Patent Literature 2 Japanese Patent Literature 2
  • Patent Literature 3 a system in which indoor flight without using a GPS is possible has been also proposed.
  • SHAOJIE SHEN ET AL "Autonomous multi-floor indoor navigation with a computationally constrained MAV", ROBOTICS AND AUTOMATION (ICRA), 2011 IEEE INTERNATIONAL CONFERENCE ON, IEEE, 9 May 2011, pages 20-25 .
  • a characteristic point (or a pattern) needs to be provided on the ground instead of the GPS in the proposal of Patent Literature 3, and there is a problem that a place where this characteristic point (or the pattern) can be installed is limited.
  • the structure such as the boiler furnace and the stack has a closed space whose inside is dark, there is a problem that it is difficult to confirm the characteristic point.
  • an indoor monitoring system for a structure that is capable of unmanned inspection which reliably obtains internal position information in a closed indoor structure such as a boiler furnace and a stack and also capable of reducing efforts, cost, inspection periods by omitting erection of scaffolding, for example.
  • the present invention has been made in view of the above-described problems, and an object thereof is to provide an indoor monitoring system and method for a structure that is capable of unmanned inspection which reliably obtains internal position information, and also capable of reducing efforts, cost, inspection periods by omitting erection of scaffolding, for example.
  • the unmanned inspection which reliably obtains the position information inside the structure, for example, the boiler furnace, the stack, or the like, and it is also possible to achieve the significant reduction of efforts, cost, inspection periods by omitting the scaffolding erection, for example.
  • the present invention is not limited by the embodiments, and further, encompasses any configuration obtained by combining the respective embodiments when there are a plurality of embodiments.
  • FIG. 1 is a schematic view of an indoor monitoring system for a structure according to a first embodiment.
  • FIG. 2 is a schematic view illustrating an aspect of inspection of a boiler furnace according to the first embodiment.
  • FIG. 3 is a block configuration diagram of an indoor monitoring system for a structure according to the first embodiment. As illustrated in FIGS.
  • the indoor monitoring system for the structure is provided with: an unmanned floating machine 11 with propellers 22, for example, as a floating means which levitates and moves the unmanned floating machine inside a closed structure 50, for example, a boiler furnace or the like by remote control; a distance measurement unit (for example, a laser scanner, an ultrasonic sensor, or the like) 12 which is mounted to the unmanned floating machine 11 and measures a distance between the unmanned floating machine 11 and an inner wall surface of the structure 50; an inertial measurement unit (IMU) which is mounted to the unmanned floating machine 11 and obtains an attitude of a body of the unmanned floating machine; an imaging unit (a still image imaging unit 13A and a video imaging unit 13B) 13 which is mounted to the unmanned floating machine 11 and images a structure (for example, piping, fitting or the like) on a wall surface side of the structure 50; an operation unit 15 which remotely controls the unmanned floating machine 11; a flight position information acquisition unit 16 which acquires current position information of the
  • the flight position information acquisition unit 16 is configured to execute: a distance measuring step (step 1: S-1) of measuring a horizontal distance information (r(t), ⁇ s ) between the unmanned floating machine 11 and the inner wall surface of the structure 50 using the distance measurement unit 12; an attitude angle acquiring step (step 2: S-2) of acquiring an attitude angle of the unmanned floating machine 11 using the inertial measurement unit; a distance correcting step (step 3: S-3) of correcting the horizontal distance information (r(t), ⁇ s ) using the attitude angle acquired in step 2; a distance acquiring step (step 4: S-4) of acquiring distances of at least two points (any two points among front (L f (t)) and left (L L (t)), front (L f (t)) and right (L R (t)), back (L B (t)) and left (L L (t)), and back (L B (t)) and right (L R (t))) on the front, back, right and left of the unmanned floating machine 11
  • the structure 50 which has a simple shape (whose cross-sectional shape is a rectangle or a circle), for example, a boiler furnace, a stack, or the like, is set as a target of inspection. Since the inside of the structure 50 is the target, provided is a system that monitors a flight position (current flight position information) of the unmanned floating machine 11 using the distance measurement unit (for example, the laser scanner, the ultrasonic sensor, or the like) 12 which does not use a GPS and the inertial measurement unit (IMU) which belongs a sensor group used for attitude control of the unmanned floating machine 11.
  • the distance measurement unit for example, the laser scanner, the ultrasonic sensor, or the like
  • IMU inertial measurement unit
  • the unmanned floating machine 11 is operated by the operation unit 15 while confirming the flight position of the unmanned floating machine 11 and an image (a damaged portion) using the monitor unit 14 of a personal computer PC in a ground station positioned outside the closed structure (boiler furnace) 50, thereby performing inspection of an inner wall of a closed space of the boiler furnace 50, as illustrated in FIG. 2 .
  • the unmanned floating machine 11 is introduced from an entrance of the boiler furnace 50 illustrated in FIG. 2 , thereafter is raised by a predetermined distance inside the boiler furnace 50, and is turned along inner surfaces of walls in the four directions by operating the operation unit 15 on the ground side. Thereafter, the unmanned floating machine 11 is raised again by a predetermined distance, and is turned along the inner wall surfaces in the four directions in the same manner. This operation is repeated until the top of the boiler furnace 50 is inspected, and then, the unmanned floating machine 11 is lowered, thereby ending the inspection.
  • a degree of damage such as a crack in the piping on the inner surface is inspected using the imaging unit. During this inspection, it is possible to confirm the flight position and the damaged portion of the closed indoor structure on the monitor unit 14 according to the present embodiment, and thus, it is possible to perform unmanned inspection that reliably obtains the internal position information.
  • the perimeter of the unmanned floating machine 11 is protected by an body guard portion 21 (a front-side guard portion 21A, a left-side guard portion 21B, a right-side guard portion 21C, and a back-side guard portion 21D), and there is provided the propeller 22 as the floating means on each upper surface of four corners of the body guard portion 21, the distance measurement unit 12 mounted at a center portion of a body 21E, the still image imaging unit 13A positioned on a part of the front-side guard portion 21A, and the video imaging unit 13B positioned on the back-side guard portion 21D via a support portion 13b, as illustrated in FIG. 1 .
  • the distance measurement unit 12 scans a predetermined angle ( ⁇ 135° in the present embodiment) and can be turned by a turning means (not illustrated).
  • any one of the still image imaging unit 13A and the video imaging unit 13B may be used as the imaging unit 13 to confirm the internal information.
  • FIG. 6 is a diagram illustrating three aspects of an attitude position of the unmanned floating machine according to the first embodiment.
  • the inertial measurement unit is a device that detects angles (or angular velocities) and accelerations in three axes governing a motion.
  • the upper stage of FIG. 6 illustrates an aspect of vertical rotation of the unmanned floating machine 11 which is turning (the pitch ( ⁇ )) where the front-side guard portion 21A (on a nose side) facing the inner wall 50a side is raised or lowered.
  • the middle stage of FIG. 6 illustrates an aspect of horizontal rotation of the body of the unmanned floating machine 11 which is turning (the yaw ( ⁇ )) where a direction of a nose is shifted right and left, and the left-side guard portion 21B and the right-side guard portion 21C swing right and left.
  • the lower stage of FIG. 6 is an aspect of rotation about an axis in a travel direction of the unmanned floating machine 11 which is turning (the roll ( ⁇ )) where the body is tilted right and left. Incidentally, the right and left of the body are based on the travel direction thereof.
  • the flight position information acquisition unit 16 is configured to obtain a real distance based on actual distance information of the distance measurement unit 12 and the attitude angle information of the inertial measurement unit (IMU). This is because there is a need for correction of the measured distance since the unmanned floating machine 11 is not always capable of flying constantly according to XY coordinates.
  • IMU inertial measurement unit
  • FIG. 7 is a flowchart of position monitoring in the horizontal direction according to the first embodiment.
  • the measurement in the horizontal direction is performed through step 1 (S-1) to step 5 (S-5).
  • an initial direction information acquiring step (S-0) of acquiring initial direction information, obtained when the unmanned floating machine 11 is positioned at the bottom inside the structure 50, is provided in the present embodiment, but this step may be omitted.
  • step 1 (S-1) it is possible to acquire the real distance information in the horizontal direction on consideration of the attitude angle at the time of measurement of the unmanned floating machine 11 by executing step 1 (S-1) to step 5 (S-5).
  • step 3 the correction of the measured distance using the attitude angle acquired in step 3 (S-3) is performed as follows.
  • a laser measurement point obtained as (r(t), ⁇ s ) is transformed into coordinates (x R ,y R ).
  • This coordinate transformation is obtained by the following Formula (1).
  • x R t y R t r t cos ⁇ s r t sin ⁇ s
  • each distance on the front, back, right and left of the unmanned floating machine 11 is obtained on the basis of the yaw angle ⁇ (t) acquired by the inertial measurement unit (IMU) in step 4 (S-4).
  • IMU inertial measurement unit
  • a current position (x(t),y(t)) is acquired from the existing horizontal cross-sectional shape using measurable distances (at least two of the front, back, right and left distance).
  • FIG. 8 is a flowchart of position monitoring in the height direction according to the first embodiment.
  • An initial direction information acquiring step (S-10) of acquiring an initial direction information uses the information obtained in the initial direction information acquiring step (S-0) of acquiring the initial direction information in the horizontal direction.
  • the measurement in the height direction is performed through the following step 11 (S-11) to step 14 (S-14).
  • a corrected measurement point (z'(t)) is obtained from the following Formula (4).
  • z ′ z ⁇ cos ⁇ ⁇ ⁇ cos ⁇ ⁇
  • FIG. 3 is a block configuration diagram of an indoor monitoring system for a structure according to the first embodiment.
  • FIG. 4 is a block configuration diagram of another indoor monitoring system for a structure according to the first embodiment.
  • the present embodiment is a case in which position information processing is executed on the unmanned floating machine 11 side.
  • the flight position information acquisition unit 16 is mounted at a predetermined portion (not illustrated) on the unmanned floating machine 11 side, and here, acquires real current position information and transmits the acquired real current position information to the ground side by a transmission unit 13a to display the information on the monitor unit 14.
  • the operation of the unmanned floating machine 11 is performed in such a manner that a reception unit 15a receives a signal from the operation unit 15 and a flight command is issued to a floating machine driving unit 19.
  • the imaging information of the imaging unit 13 (the still image imaging unit 13A and the video imaging unit 13B) 13 is transmitted to the ground side by the transmission unit 13a at the same time and displayed on the monitor unit 14 in the present embodiment.
  • FIG. 4 Another example illustrated in FIG. 4 is a case in which the position information processing is performed on a controller terminal side of the personal computer PC on the ground.
  • the flight position information acquisition unit 16 is mounted to the controller terminal of the PC on the ground side (base station), and information (signal) of the distance measurement unit 12 and information (signal) of the inertial measurement unit (IMU) are transmitted to the ground side by the transmission unit 13a. Further, the received information is processed by the flight position information acquisition unit 16 to acquire real current position information, and this acquired current position information is displayed on the monitor unit 14.
  • the present invention is not limited thereto, and for example, may be configured such that the imaging information is temporarily stored in a memory unit of the imaging unit on the unmanned floating machine 11 side, the information is transmitted to the ground station side after ending measurement, and the imaging information and the position information are processed to match each other.
  • the unmanned inspection which reliably obtains the position information inside the structure 50, for example, the boiler furnace, the stack, or the like, and it is possible to achieve the significant reduction of efforts, cost, inspection periods by omitting erection of scaffolding, for example, according to the present embodiment.
  • the present invention is not limited thereto, and the accuracy in position measurement may be improved based on measurement information at multiple points.
  • multiple points are extracted and averaged based on the scan angle in the distance measurement unit 12 to obtain each distance in the calculation of distances in the horizontal direction and the height direction. Further, when more than half of the multiple points is abnormal for distance measurement or unmeasurable, such points are not used for the position monitoring.

Landscapes

  • Engineering & Computer Science (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Automation & Control Theory (AREA)
  • Analytical Chemistry (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Chemical & Material Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Biochemistry (AREA)
  • General Health & Medical Sciences (AREA)
  • Immunology (AREA)
  • Pathology (AREA)
  • Electromagnetism (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
  • Studio Devices (AREA)
  • Length Measuring Devices By Optical Means (AREA)
  • Closed-Circuit Television Systems (AREA)

Description

    Field
  • The present invention relates to an indoor monitoring system and method for a structure.
  • Background
  • For example, a boiler furnace used at a thermal power plant needs to be opened during construction and periodically after starting operation so that a worker enters the inside to conduct maintenance inspection. During this maintenance inspection, it is necessary to define a portion to be inspected, but it is difficult to accurately grasp the portion to be inspected visually because the capacity of the boiler furnace is large. Thus, a height position and a lateral position of the portion to be inspected have been conventionally measured and marked using a measuring tape or the like to grasp where the worker is or a maintenance inspection position, but such a method requires not only erection of scaffolding for the worker and installation of a gondola but also a lot of efforts, cost, and inspection periods.
  • Thus, a technique has been conventionally proposed to clean up the inside of a structure such as a stack using an unmanned inspection apparatus (Patent Literature 1). However, this proposal also requires a cradle to install a wire, and efforts, cost and inspection periods are required for the preparation thereof.
  • In addition, application of an unmanned inspection technique, which does not require erection of scaffolding using an unmanned aircraft and GPS (Global Positioning System), has been proposed for an outdoor structure (Patent Literature 2).
  • However, since electric waves from satellites do not reach indoor structures such as the inside of a boiler and a stack, it is difficult to obtain a flight position using the GPS and to stably maneuver the unmanned aircraft. Thus it is difficult to use the existing inspection technique using the unmanned aircraft.
  • In this regard, a system in which indoor flight without using a GPS is possible has been also proposed (Patent Literature 3).
  • Citation List Patent Literature
    • Patent Literature 1: Japanese Laid-open Patent Publication No. 6-73922
    • Patent Literature 2: Japanese Patent Application National Publication (Laid-Open) No. 2011-530692
    • Patent Literature 3: EP 1901153 A
  • Further prior art is SHAOJIE SHEN ET AL: "Autonomous multi-floor indoor navigation with a computationally constrained MAV", ROBOTICS AND AUTOMATION (ICRA), 2011 IEEE INTERNATIONAL CONFERENCE ON, IEEE, 9 May 2011, pages 20-25.
  • Summary Technical Problem
  • However, a characteristic point (or a pattern) needs to be provided on the ground instead of the GPS in the proposal of Patent Literature 3, and there is a problem that a place where this characteristic point (or the pattern) can be installed is limited. In addition, since the structure such as the boiler furnace and the stack has a closed space whose inside is dark, there is a problem that it is difficult to confirm the characteristic point.
  • Accordingly, there has been a request for emergence of an indoor monitoring system for a structure that is capable of unmanned inspection which reliably obtains internal position information in a closed indoor structure such as a boiler furnace and a stack and also capable of reducing efforts, cost, inspection periods by omitting erection of scaffolding, for example.
  • The present invention has been made in view of the above-described problems, and an object thereof is to provide an indoor monitoring system and method for a structure that is capable of unmanned inspection which reliably obtains internal position information, and also capable of reducing efforts, cost, inspection periods by omitting erection of scaffolding, for example.
  • Solution to Problem
  • Accordingly, there is provided an indoor monitoring system as set out in independent claim 1, and an indoor monitoring method as set out in independent claim 10. Advantageous developments are defined in the dependent claims.
  • Advantageous Effects of Invention
  • According to the present invention, it is possible to perform the unmanned inspection which reliably obtains the position information inside the structure, for example, the boiler furnace, the stack, or the like, and it is also possible to achieve the significant reduction of efforts, cost, inspection periods by omitting the scaffolding erection, for example.
  • Brief Description of Drawings
    • FIG. 1 is a schematic view of an unmanned floating machine according to a first embodiment.
    • FIG. 2 is a schematic view illustrating an aspect of inspection of a boiler furnace according to the first embodiment.
    • FIG. 3 is a block configuration diagram of an indoor monitoring system for a structure according to the first embodiment.
    • FIG. 4 is a block configuration diagram of another indoor monitoring system for a structure according to the first embodiment.
    • FIG. 5 is a diagram illustrating an example of a scan range in a case where a laser scanner is used as a distance measurement unit according to the first embodiment.
    • FIG. 6 is a diagram illustrating three aspects of an attitude position of the unmanned floating machine according to the first embodiment.
    • FIG. 7 is a flowchart of position monitoring in the horizontal direction according to the first embodiment.
    • FIG. 8 is a flowchart of position monitoring in the height direction according to the first embodiment.
    • FIG. 9 is a diagram illustrating an example of acquisition of a current position in the horizontal direction according to the first embodiment.
    Description of Embodiments
  • Hereinafter, preferred embodiments of the present invention will be described in detail with reference to the drawings. Incidentally, the present invention is not limited by the embodiments, and further, encompasses any configuration obtained by combining the respective embodiments when there are a plurality of embodiments.
  • First Embodiment
  • FIG. 1 is a schematic view of an indoor monitoring system for a structure according to a first embodiment. FIG. 2 is a schematic view illustrating an aspect of inspection of a boiler furnace according to the first embodiment. FIG. 3 is a block configuration diagram of an indoor monitoring system for a structure according to the first embodiment. As illustrated in FIGS. 1 to 3, the indoor monitoring system for the structure according to the present embodiment is provided with: an unmanned floating machine 11 with propellers 22, for example, as a floating means which levitates and moves the unmanned floating machine inside a closed structure 50, for example, a boiler furnace or the like by remote control; a distance measurement unit (for example, a laser scanner, an ultrasonic sensor, or the like) 12 which is mounted to the unmanned floating machine 11 and measures a distance between the unmanned floating machine 11 and an inner wall surface of the structure 50; an inertial measurement unit (IMU) which is mounted to the unmanned floating machine 11 and obtains an attitude of a body of the unmanned floating machine; an imaging unit (a still image imaging unit 13A and a video imaging unit 13B) 13 which is mounted to the unmanned floating machine 11 and images a structure (for example, piping, fitting or the like) on a wall surface side of the structure 50; an operation unit 15 which remotely controls the unmanned floating machine 11; a flight position information acquisition unit 16 which acquires current position information of the unmanned floating machine 11 based on information (signal) of the distance measurement unit 12 and information (signal) of the inertial measurement unit; and a monitor unit 14 which displays image information from the imaging unit 13 and position information from the flight position information acquisition unit 16. Incidentally, 12a represents a laser light emitting portion.
  • Further, the flight position information acquisition unit 16 is configured to execute: a distance measuring step (step 1: S-1) of measuring a horizontal distance information (r(t),αs) between the unmanned floating machine 11 and the inner wall surface of the structure 50 using the distance measurement unit 12; an attitude angle acquiring step (step 2: S-2) of acquiring an attitude angle of the unmanned floating machine 11 using the inertial measurement unit; a distance correcting step (step 3: S-3) of correcting the horizontal distance information (r(t),αs) using the attitude angle acquired in step 2; a distance acquiring step (step 4: S-4) of acquiring distances of at least two points (any two points among front (Lf(t)) and left (LL(t)), front (Lf(t)) and right (LR(t)), back (LB(t)) and left (LL(t)), and back (LB(t)) and right (LR(t))) on the front, back, right and left of the unmanned floating machine 11 on the basis of a yaw angle acquired by the inertial measurement unit; and a horizontal-direction current position information acquiring step (step 5: S-5) of acquiring current position information in the horizontal direction from existing horizontal cross-sectional shape information of the structure 50.
  • In the present embodiment, the structure 50, which has a simple shape (whose cross-sectional shape is a rectangle or a circle), for example, a boiler furnace, a stack, or the like, is set as a target of inspection. Since the inside of the structure 50 is the target, provided is a system that monitors a flight position (current flight position information) of the unmanned floating machine 11 using the distance measurement unit (for example, the laser scanner, the ultrasonic sensor, or the like) 12 which does not use a GPS and the inertial measurement unit (IMU) which belongs a sensor group used for attitude control of the unmanned floating machine 11.
  • In the present embodiment, the unmanned floating machine 11 is operated by the operation unit 15 while confirming the flight position of the unmanned floating machine 11 and an image (a damaged portion) using the monitor unit 14 of a personal computer PC in a ground station positioned outside the closed structure (boiler furnace) 50, thereby performing inspection of an inner wall of a closed space of the boiler furnace 50, as illustrated in FIG. 2.
  • During the inspection, the unmanned floating machine 11 is introduced from an entrance of the boiler furnace 50 illustrated in FIG. 2, thereafter is raised by a predetermined distance inside the boiler furnace 50, and is turned along inner surfaces of walls in the four directions by operating the operation unit 15 on the ground side. Thereafter, the unmanned floating machine 11 is raised again by a predetermined distance, and is turned along the inner wall surfaces in the four directions in the same manner. This operation is repeated until the top of the boiler furnace 50 is inspected, and then, the unmanned floating machine 11 is lowered, thereby ending the inspection.
  • A degree of damage such as a crack in the piping on the inner surface is inspected using the imaging unit. During this inspection, it is possible to confirm the flight position and the damaged portion of the closed indoor structure on the monitor unit 14 according to the present embodiment, and thus, it is possible to perform unmanned inspection that reliably obtains the internal position information.
  • The perimeter of the unmanned floating machine 11 is protected by an body guard portion 21 (a front-side guard portion 21A, a left-side guard portion 21B, a right-side guard portion 21C, and a back-side guard portion 21D), and there is provided the propeller 22 as the floating means on each upper surface of four corners of the body guard portion 21, the distance measurement unit 12 mounted at a center portion of a body 21E, the still image imaging unit 13A positioned on a part of the front-side guard portion 21A, and the video imaging unit 13B positioned on the back-side guard portion 21D via a support portion 13b, as illustrated in FIG. 1. Incidentally, the distance measurement unit 12 scans a predetermined angle (±135° in the present embodiment) and can be turned by a turning means (not illustrated).
  • Here, any one of the still image imaging unit 13A and the video imaging unit 13B may be used as the imaging unit 13 to confirm the internal information.
  • Hereinafter, a description will be given regarding procedure of position monitoring in a case where the laser scanner is used as the distance measurement unit 12 in the present embodiment.
  • <Monitoring in Horizontal Direction>
    1. (1) First, a distance (r(t),αs) is acquired by the distance measurement unit 12 to implement monitoring in the horizontal direction.
      Here, FIG. 5 illustrates an example of a scan range of the laser scanner. In the present embodiment, a scanner-type range sensor, "UTM-30LX (product name)" manufactured by HOKUYO AUTOMATIC CO., LTD. is used.
      As illustrated in FIG. 5, this scanner-type range sensor is a two-dimensional scanning-type optical distance sensor, which measures a distance to an object to be inspected while performing scanning with laser light, and a scan angle is ±135° with 0° as the center thereof.
      In FIG. 5, a distance (r) is an actually measured distance obtained when measurement is performed up to an inner wall 50a from the laser scanner of the distance measurement unit 12, and α is an angle at the measured scanning step thereof. A scanning step (s) for measurement in this device is set to every 0.25°.
    2. (2) Next, attitude angles including a pitch angle (θ(t)), a yaw angle (ψ(t)), a roll angle (ϕ(t)) of the unmanned floating machine 11 are acquired by the inertial measurement unit (IMU).
  • FIG. 6 is a diagram illustrating three aspects of an attitude position of the unmanned floating machine according to the first embodiment.
  • The inertial measurement unit (IMU) is a device that detects angles (or angular velocities) and accelerations in three axes governing a motion.
  • Here, the upper stage of FIG. 6 illustrates an aspect of vertical rotation of the unmanned floating machine 11 which is turning (the pitch (θ)) where the front-side guard portion 21A (on a nose side) facing the inner wall 50a side is raised or lowered. The middle stage of FIG. 6 illustrates an aspect of horizontal rotation of the body of the unmanned floating machine 11 which is turning (the yaw (ψ)) where a direction of a nose is shifted right and left, and the left-side guard portion 21B and the right-side guard portion 21C swing right and left. The lower stage of FIG. 6 is an aspect of rotation about an axis in a travel direction of the unmanned floating machine 11 which is turning (the roll (ϕ)) where the body is tilted right and left. Incidentally, the right and left of the body are based on the travel direction thereof.
  • Next, a position monitoring measuring step will be described with reference to FIG. 3.
  • The flight position information acquisition unit 16 is configured to obtain a real distance based on actual distance information of the distance measurement unit 12 and the attitude angle information of the inertial measurement unit (IMU). This is because there is a need for correction of the measured distance since the unmanned floating machine 11 is not always capable of flying constantly according to XY coordinates.
  • FIG. 7 is a flowchart of position monitoring in the horizontal direction according to the first embodiment.
  • The measurement in the horizontal direction is performed through step 1 (S-1) to step 5 (S-5).
  • Prior to this measurement, an initial direction information acquiring step (S-0) of acquiring initial direction information, obtained when the unmanned floating machine 11 is positioned at the bottom inside the structure 50, is provided in the present embodiment, but this step may be omitted.
    1. 1) Step 1 is the horizontal-direction distance measuring step (S-1) of measuring the horizontal distance information (r(t),αs) between the unmanned floating machine 11 and the inner wall 50a of the structure 50 using the distance measurement unit 12.
    2. 2) Step 2 is the attitude angle acquiring step (S-2) of acquiring the attitude angle of the unmanned floating machine 11 using the inertial measurement unit (IMU).
    3. 3) Step 3 is a horizontal-direction distance correcting step (S-3) of correcting the horizontal distance information (r(t),αs) using the attitude angle acquired in step 2 (S-2).
    4. 4) Step 4 is a horizontal-direction distance acquiring step (S-4) of acquiring distances of at least two points (any two points among front (LF(t)) and left (LL(t)), front (LF(t)) and right (LR(t)), back (LB(t)) and left (LL(t)), and back (LB(t)) and right (LR(t))) on the front, back, right and left of the unmanned floating machine 11 on the basis of the yaw angle (ψ) acquired by the inertial measurement unit (IMU), as illustrated in FIG. 9.
    5. 5) Step 5 is the horizontal-direction current position information acquiring step (S-5) of acquiring the current position information in the horizontal direction from the existing horizontal cross-sectional shape information of the structure 50.
  • It is possible to acquire the real distance information in the horizontal direction on consideration of the attitude angle at the time of measurement of the unmanned floating machine 11 by executing step 1 (S-1) to step 5 (S-5).
  • Here, the correction of the measured distance using the attitude angle acquired in step 3 (S-3) is performed as follows.
  • A laser measurement point obtained as (r(t),αs) is transformed into coordinates (xR,yR). This coordinate transformation is obtained by the following Formula (1). x R t y R t = r t cos α s r t sin α s
    Figure imgb0001
  • Transformation of a corrected measurement point (x'(t),y'(t)) into a rotation coordinate system is obtained by the following Formula (2). x t y t z t 1 = 1 0 0 0 0 cos θ t sin θ t 0 0 sin θ t cos θ t 0 0 0 0 1 cos φ t 0 sin φ t 0 0 1 0 0 sin φ t 0 cos φ t 0 0 0 0 1 x R t y R t z 1
    Figure imgb0002
  • A value obtained from Formula (2) is transformed into a coordinate system (r,α) of laser measurement. This coordinate transformation is obtained by the following Formula (3). r t a t = x 2 + y 2 arctan y / x
    Figure imgb0003
  • Next, each distance on the front, back, right and left of the unmanned floating machine 11 is obtained on the basis of the yaw angle ψ(t) acquired by the inertial measurement unit (IMU) in step 4 (S-4). However, when a scan angle is out of a predetermined scan range, the obtained data is not employed as a distance from a wall side.
    • Scan angle data where a scan angle α1 = ψ(t) is set as a front distance LF(t).
    • Scan angle where a scan angle α2 = ψ(t) - 90° is set as a left distance LL(t).
    • Scan angle where a scan angle α3 = ψ(t) + 90° is set as a right distance LR(t).
    • Scan angle where a scan angle α4 = ψ(t) + 180° is set as a back distance LB(t).
  • In the final step 5 (S-5), a current position (x(t),y(t)) is acquired from the existing horizontal cross-sectional shape using measurable distances (at least two of the front, back, right and left distance).
  • Accordingly, it is possible to acquire the real current position, and it is possible to confirm the imaging information imaged in this current position and the position information by the monitor unit 14.
  • When this measurement of the position information is performed every time when the unmanned floating machine 11 travels, it is possible to reliably obtaining the position information continuously.
  • <Monitoring in Height Direction>
  • FIG. 8 is a flowchart of position monitoring in the height direction according to the first embodiment.
  • An initial direction information acquiring step (S-10) of acquiring an initial direction information uses the information obtained in the initial direction information acquiring step (S-0) of acquiring the initial direction information in the horizontal direction.
  • The measurement in the height direction is performed through the following step 11 (S-11) to step 14 (S-14).
    • 6) Step 11 is a height-direction distance measuring step (S-11) of measuring the distance information (LD(t),αS) between the unmanned floating machine 11 and the structure 50 on the lower side in the height direction using the distance measurement unit 12.
      Here, the measurement in the height direction using laser light is performed using a reflective optical system such as a mirror (not illustrated). When an irradiation distance of the laser light does not reach as the unmanned floating machine 11 is raised, distance information (LU(t),αs) on the upper side may be measured by causing the laser light to be reflected to the upper side.
    • 7) Step 12 is an attitude angle acquiring step (S-12) of acquiring an attitude angle of the unmanned floating machine 11 using the inertial measurement unit (IMU).
    • 8) Step 13 is a height-direction distance correcting step (S-13) of correcting the distance information (LD(t)) in the height direction using an attitude angle (ϕ(t),θ(t)) acquired in step 12 (S-12).
    • 9) Step 14 is a height-direction current position information acquiring step (S-14) of acquiring current position information in the height direction from existing vertical cross-sectional shape information of the structure 50.
  • In the correction in step 13 (S-13), a corrected measurement point (z'(t)) is obtained from the following Formula (4). z = z cos α cos β
    Figure imgb0004
  • Therefore, it is possible to transform the actually measured distances in the horizontal direction and the height direction into the real distances and to reliably acquire the position information.
  • As a result, it is possible to perform the inspection that reliably obtains the measurement position using the unmanned floating machine inside the structure 50 where it is difficult to use the GPS. As a result, it is unnecessary to erect scaffolding inside the structure 50 as in the related art, and it is possible to significantly reduce efforts, cost, and inspection periods for internal inspection.
  • FIG. 3 is a block configuration diagram of an indoor monitoring system for a structure according to the first embodiment. FIG. 4 is a block configuration diagram of another indoor monitoring system for a structure according to the first embodiment.
  • As illustrated in FIG. 3, the present embodiment is a case in which position information processing is executed on the unmanned floating machine 11 side.
  • In the present embodiment, the flight position information acquisition unit 16 is mounted at a predetermined portion (not illustrated) on the unmanned floating machine 11 side, and here, acquires real current position information and transmits the acquired real current position information to the ground side by a transmission unit 13a to display the information on the monitor unit 14.
  • Incidentally, the operation of the unmanned floating machine 11 is performed in such a manner that a reception unit 15a receives a signal from the operation unit 15 and a flight command is issued to a floating machine driving unit 19.
  • In addition, the imaging information of the imaging unit 13 (the still image imaging unit 13A and the video imaging unit 13B) 13 is transmitted to the ground side by the transmission unit 13a at the same time and displayed on the monitor unit 14 in the present embodiment.
  • With respect to this, another example illustrated in FIG. 4 is a case in which the position information processing is performed on a controller terminal side of the personal computer PC on the ground.
  • In this example, the flight position information acquisition unit 16 is mounted to the controller terminal of the PC on the ground side (base station), and information (signal) of the distance measurement unit 12 and information (signal) of the inertial measurement unit (IMU) are transmitted to the ground side by the transmission unit 13a. Further, the received information is processed by the flight position information acquisition unit 16 to acquire real current position information, and this acquired current position information is displayed on the monitor unit 14.
  • Although the imaging information imaged by the imaging unit 13 is transmitted by the transmission unit 13a in the present embodiment, the present invention is not limited thereto, and for example, may be configured such that the imaging information is temporarily stored in a memory unit of the imaging unit on the unmanned floating machine 11 side, the information is transmitted to the ground station side after ending measurement, and the imaging information and the position information are processed to match each other.
  • As described above, it is possible to perform the unmanned inspection which reliably obtains the position information inside the structure 50, for example, the boiler furnace, the stack, or the like, and it is possible to achieve the significant reduction of efforts, cost, inspection periods by omitting erection of scaffolding, for example, according to the present embodiment.
  • Second Embodiment
  • Although the measurement of the distance measurement unit 12 is performed to obtain the information of the single point in the first embodiment, the present invention is not limited thereto, and the accuracy in position measurement may be improved based on measurement information at multiple points.
  • That is, multiple points are extracted and averaged based on the scan angle in the distance measurement unit 12 to obtain each distance in the calculation of distances in the horizontal direction and the height direction. Further, when more than half of the multiple points is abnormal for distance measurement or unmeasurable, such points are not used for the position monitoring.
  • As a result, it is possible to reduce influence of a distance acquisition error.
  • Reference Signs List
    • 11 UNMANNED FLOATING MACHINE
    • 12 DISTANCE MEASUREMENT UNIT (LASER SCANNER)
    • 13 IMAGING UNIT (STILL IMAGE IMAGING UNIT 13A, VIDEO IMAGING UNIT 13B)
    • 14 MONITOR UNIT
    • 15 OPERATION UNIT
    • 16 FLIGHT POSITION INFORMATION ACQUISITION UNIT
    • 21 BODY GUARD PORTION
    • 22 PROPELLER
    • 50 STRUCTURE

Claims (17)

  1. An indoor monitoring system for a structure (50) comprising:
    an unmanned floating machine (11) including a floating means (22) which levitates the unmanned floating machine (11) inside (50) the structure by remote control, the structure being a boiler furnace and a stack;
    a distance measurement unit (12) which is mounted to the unmanned floating machine (11) and configured to measure a distance between the unmanned floating machine (11) and an inner wall surface of the structure (50);
    an inertial measurement unit (IMU) which is mounted to the unmanned floating machine (11) and configured to obtain an attitude of a body of the unmanned floating machine (11);
    an imaging unit (13) which is mounted to the unmanned floating machine (11) and configured to image a structural body on the wall surface side of the structure (50);
    an operation unit (15) which is configured to remotely control the unmanned floating machine (11);
    a flight position information acquisition unit (16) which is configured to acquire current position information of the unmanned floating machine (11) based on information from the distance measurement unit (12) and information from the inertial measurement unit (IMU); and
    a monitor unit (14) which is configured to display image information from the imaging unit (13) and position information from the flight position information acquisition unit (16),
    wherein the flight position information acquisition unit (16) is further configured to execute
    an initial direction information acquiring step of acquiring initial direction information, obtained when the unmanned floating machine (11) is positioned at the bottom inside the structure (50),
    a horizontal-direction distance measuring step of measuring hοrizontal distance in formation between the unmanned floating machine (11) and the inner wall surface of the structure (50) using the distance measurement unit (12),
    an attitude angle acquiring step of acquiring an attitude angle of the unmanned floating machine (11) using the inertial measurement unit (IMU),
    a horizontal-direction distance correcting step of correcting the horizontal distance information using the attitude angle acquired in the attitude angle acquiring step,
    a horizontal-direction distance acquiring step of acquiring distances between the inner wall surface of the structure (50) and the unmanned floating machine (11) in at least two different horizontal directions around the unmanned floating machine (11) based on a yaw angle (ψ(t)) acquired by the inertial measurement unit (IMU), and
    a horizontal-direction current position information acquiring step of acquiring current position information in a horizontal direction from existing horizontal cross-sectional shape information of the structure (50),
    wherein
    when the unmanned floating machine (11) is introduced from an entrance of the structure, the operation unit (15) is configured to remotely control the unmanned floating machine (11),
    (i) to raise by a predetermined distance inside the structure,
    (ii) to turn along inner surfaces of walls in the four directions,
    (iii) to raise again by the predetermined distance, and
    (iv) to turn along the inner surfaces of the walls in the four directions in the same manner,
    to repeat the operations (i) to (iv) until the unmanned floating machine (11) reaches a top of the structure, and
    to lower the unmanned floating machine (11).
  2. The indoor monitoring system for a structure (50) according to claim 1, wherein
    the flight position information acquisition unit (16) is further configured to execute
    a height-direction distance measuring step of measuring distance information in any height direction between the unmanned floating machine (11) and a top or a bottom of the structure (50) using the distance measurement unit (12),
    the attitude angle acquiring step of acquiring the attitude angle of the unmanned floating machine (11) using the inertial measurement unit (IMU),
    a height-direction distance correcting step of correcting the distance information in the height direction using the attitude angle acquired in the attitude angle acquiring step, and
    a height-direction current position information acquiring step of acquiring current position information in the height direction from existing vertical cross-sectional shape information of the structure (50).
  3. The indoor monitoring system for a structure (50) according to claim 1, wherein
    multiple points are measured in the horizontal-direction distance measuring step and an averaged distance is used as the horizontal distance information.
  4. The indoor monitoring system for a structure (50) according to claim 2, wherein
    multiple points are measured in the height-direction distance measuring step and an averaged distance is used as the distance information in the height direction.
  5. The indoor monitoring system for a structure (50) according to claim 1 or 2, wherein
    the flight position information acquisition unit (16) is mounted to the unmanned floating machine (11) and transmits the acquired current position information to a ground side by a transmission unit (13a) to display the current position information on the monitor unit (14).
  6. The indoor monitoring system for a structure (50) according to claim 1 or 2, wherein
    the flight position information acquisition unit (16) is mounted to a controller terminal on a ground side, and
    the information from the distance measurement unit (12) and the information from the inertial measurement unit (IMU) are transmitted to the ground side by a transmission unit (13a) and processed by the flight position information acquisition unit (16), and the current position information is displayed on the monitor unit (14).
  7. The indoor monitoring system for a structure (50) according to claim 1 or 2, wherein
    the imaging unit (13) is any one or both of a still image imaging unit and a video imaging unit.
  8. The indoor monitoring system for a structure (50) according to claim 1 or 2, wherein
    a guard portion (21) is provided around the unmanned floating machine (11).
  9. The indoor monitoring system for a structure (50) according to claim 1, wherein
    the flight position information acquisition unit (16) is configured to measure a plurality of horizontal distance information, and not to use the plurality of horizontal distance information when more than half of the plurality of measured points is abnormal for distance measurement.
  10. An indoor monitoring method for a structure (50), the structure being a boiler furnace and a stack, which uses an unmanned floating machine (11) including a floating means (22) which levitates the unmanned floating machine (11) inside the structure (50) by remote control, the method comprising:
    a distance measurement step executed in the unmanned floating machine (11) and measuring a distance between the unmanned floating machine (11) and an inner wall surface of the structure (50);
    an inertial measurement step executed in the unmanned floating machine and obtaining an attitude of a body of the unmanned floating machine;
    an imaging step executed in the unmanned floating machine (11) and imaging a structural body on the wall surface side of the structure (50);
    an operation step of remotely controlling the unmanned floating machine (11);
    a flight position information acquisition step of acquiring current position information of the unmanned floating machine (11) based on information from the distance measurement step and information from the inertial measurement step; and
    a monitor displaying step of displaying image information from the imaging step and position information from the flight position information acquisition step,
    wherein in the flight position information acquisition step,
    an initial direction information acquiring step of acquiring initial direction information, obtained when the unmanned floating machine (11) is positioned at the bottom inside the structure (50),
    a horizontal-direction distance measuring step of measuring horizontal distance information between the unmanned floating machine (11) and the inner wall surface of the structure (50) is executed using the distance measurement step,
    an attitude angle acquiring step of acquiring an attitude angle of the unmanned floating machine (11) is executed using the inertial measurement step,
    a horizontal-direction distance correcting step of correcting the horizontal distance information using the attitude angle acquired in the attitude angle acquiring step is executed,
    a horizontal-direction distance acquiring step of acquiring distances between the inner wall surface of the structure (50) and the unmanned floating machine (11) in at least two different horizontal directions around the unmanned floating machine (11) based on a yaw angle (ψ(t)) acquired in the inertial measurement step is executed,
    a horizontal-direction current position information acquiring step of acquiring current position information in a horizontal direction from existing horizontal cross-sectional shape information of the structure (50) is executed, and
    in the operating step
    an introducing step causing the unmanned floating machine (11) introduced from an entrance of the structure,
    (i) a first raising step causing the unmanned floating machine (11) to raise by a predetermined distance inside the structure,
    (ii) a first turning step causing the unmanned floating machine (11) to turn along inner surfaces of walls in the four directions,
    (iii) a second raising step causing the unmanned floating machine (11) to raise again by the predetermined distance, and
    (iv) a second turning step causing the unmanned floating machine (11) to turn along the inner surfaces of the wall in the four directions in the same manner,
    a repeating step repeating the operations (i) to (iv) until the unmanned floating machine (11) reaches a top of the structure, and
    a lowering step causing the unmanned floating machine (11) to lower.
  11. The indoor monitoring method for a structure (50) according to claim 10, wherein
    in the flight position information acquisition step,
    a height-direction distance measuring step of measuring distance information in any height direction between the unmanned floating machine (11) and a top or a bottom of the structure is executed using the distance measurement step,
    the attitude angle acquiring step of acquiring the attitude angle of the unmanned floating machine (11) is executed using the inertial measurement step,
    a height-direction distance correcting step of correcting the distance information in the height direction using the attitude angle acquired in the attitude angle acquiring step is executed, and
    a height-direction current position information acquiring step of acquiring current position information in the height direction from existing vertical cross-sectional shape information of the structure (50) is executed.
  12. The indoor monitoring method for a structure (50) according to claim 10, wherein
    multiple points are measured in the horizontal-direction distance measuring step and an averaged distance is used as the horizontal distance information.
  13. The indoor monitoring method for a structure (50) according to claim 11, wherein
    multiple points are measured in the height-direction distance measuring step and an averaged distance is used as the distance information in the height direction.
  14. The indoor monitoring method for a structure (50) according to claim 10 or 11, wherein
    the flight position information acquisition step is processed by the unmanned floating machine side to transmit the acquired current position information to a ground side to be monitor-displayed.
  15. The indoor monitoring method for a structure (50) according to claim 10 or 11, wherein
    the flight position information acquisition step is processed by a ground side, and
    the information in the distance measurement step and the information in the inertial measurement step are transmitted to the ground side and processed in the flight position information acquisition step, and the current position information is monitor-displayed.
  16. The indoor monitoring method for a structure (50) according to claim 10 or 11, wherein
    the imaging step is any one or both of a still image imaging step and a video imaging step.
  17. The indoor monitoring method for a structure (50) according to claim 10, wherein
    the flight position information acquisition step is configured to measure a plurality of horizontal distance information, and not to use the plurality of horizontal distance information when more than half of the plurality of measured points is abnormal for distance measurement.
EP15814504.5A 2014-07-02 2015-01-20 Indoor monitoring system and method for structure Active EP3147743B1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2014136868A JP6486024B2 (en) 2014-07-02 2014-07-02 Indoor monitoring system and method for structure
PCT/JP2015/051360 WO2016002236A1 (en) 2014-07-02 2015-01-20 Indoor monitoring system and method for structure

Publications (3)

Publication Number Publication Date
EP3147743A1 EP3147743A1 (en) 2017-03-29
EP3147743A4 EP3147743A4 (en) 2017-05-31
EP3147743B1 true EP3147743B1 (en) 2019-04-03

Family

ID=55018806

Family Applications (1)

Application Number Title Priority Date Filing Date
EP15814504.5A Active EP3147743B1 (en) 2014-07-02 2015-01-20 Indoor monitoring system and method for structure

Country Status (6)

Country Link
US (1) US10359778B2 (en)
EP (1) EP3147743B1 (en)
JP (1) JP6486024B2 (en)
CN (1) CN106662880B (en)
TW (1) TWI578132B (en)
WO (1) WO2016002236A1 (en)

Families Citing this family (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017138049A1 (en) * 2016-02-10 2017-08-17 パナソニックIpマネジメント株式会社 Flying body and control system therefor
JP6860003B2 (en) * 2016-03-28 2021-04-14 日本電気株式会社 Unmanned aerial vehicle control system, unmanned aerial vehicle control method and inspection equipment
JP6710114B2 (en) * 2016-06-21 2020-06-17 株式会社日立製作所 Pipeline inspection vehicle and pipeline inspection system using it
JP6697821B2 (en) * 2016-07-21 2020-05-27 三菱重工機械システム株式会社 Image processing apparatus and image processing method
JP6746137B2 (en) * 2016-11-02 2020-08-26 株式会社プロドローン Unmanned aerial vehicle
CN106500611A (en) * 2016-11-25 2017-03-15 重庆科技学院 Range finding instrument auxiliary device
JP2018116443A (en) * 2017-01-18 2018-07-26 住友重機械工業株式会社 Inspection system
CN110366670B (en) * 2017-03-02 2021-10-26 深圳市大疆创新科技有限公司 Three-dimensional shape estimation method, flight vehicle, mobile platform, program, and recording medium
CN110392819B (en) * 2017-03-12 2022-02-01 株式会社尼罗沃克 Unmanned aerial vehicle for measuring water depth of farm
JP6822267B2 (en) * 2017-03-28 2021-01-27 富士通株式会社 Flyer and how to use the flyer
JP6729879B2 (en) * 2017-04-06 2020-07-29 株式会社自律制御システム研究所 Unmanned aerial vehicle and method of using the same
TWI662505B (en) * 2017-07-28 2019-06-11 中華大學 Fungus growth monitoring system
CN109561275A (en) * 2017-09-27 2019-04-02 湖南航天远望科技有限公司 A kind of area monitoring method and regional monitoring system based on circular scanning
CN111164012A (en) * 2017-09-28 2020-05-15 株式会社OPTiM Maintenance device control system, maintenance device control method, and program
JP7063578B2 (en) * 2017-11-09 2022-05-09 株式会社Soken Flight equipment
JP6475377B1 (en) * 2018-03-14 2019-02-27 株式会社サンメイ Inspection system for inspecting chimney interior and method for inspecting chimney interior
JP7033969B2 (en) * 2018-03-19 2022-03-11 三菱重工業株式会社 Unmanned flying object
CN109031312B (en) * 2018-04-26 2023-08-22 中国计量大学 Flight platform positioning device and positioning method suitable for chimney internal operation
JP6505927B1 (en) * 2018-07-24 2019-04-24 ミスギ工業株式会社 Inspection method using unmanned small-sized flying object and unmanned small-sized flying object used therefor
JP2020094780A (en) * 2018-12-14 2020-06-18 三菱重工機械システム株式会社 Chimney shaft inner surface inspection system
SG11202100153SA (en) * 2018-12-27 2021-02-25 Performance Rotors Pte Ltd Drone for surface defects inspection
JP7213104B2 (en) * 2019-02-27 2023-01-26 三菱重工業株式会社 Unmanned aerial vehicles and inspection methods
JP6733925B2 (en) * 2019-03-27 2020-08-05 ミスギ工業株式会社 Inspection method using unmanned small air vehicle and unmanned small air vehicle used for the inspection method
JP6737521B2 (en) * 2019-08-27 2020-08-12 株式会社センシンロボティクス Height position acquisition system and imaging system, height position acquisition method, and imaging method
JP6604681B1 (en) 2019-09-11 2019-11-13 株式会社Liberaware Dimension display system and dimension display method
WO2021084589A1 (en) * 2019-10-28 2021-05-06 株式会社センシンロボティクス Aerial vehicle, inspection method, and inspection system
JP6777804B1 (en) * 2019-11-22 2020-10-28 三菱パワー株式会社 Boiler internal inspection method
CN112729312A (en) * 2020-12-25 2021-04-30 云南电网有限责任公司昆明供电局 Unmanned aerial vehicle inspection method for high-voltage chamber of transformer substation
CN113483616B (en) * 2021-07-06 2023-03-14 重庆市地质矿产勘查开发局208水文地质工程地质队(重庆市地质灾害防治工程勘查设计院) Portable rapid assembling and disassembling accurate monitor for surface cracks

Family Cites Families (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3153832B2 (en) 1992-08-26 2001-04-09 三菱重工業株式会社 Chimney cleaning equipment
JP2001209426A (en) * 2000-01-26 2001-08-03 Nippon Telegr & Teleph Corp <Ntt> Mobile body controller
JP2004211995A (en) * 2003-01-07 2004-07-29 Mitsubishi Heavy Ind Ltd Inspection system of sealed space
CA2484422A1 (en) 2004-10-08 2006-04-08 Furgro Airborne Surveys Unmanned airborne vehicle for geophysical surveying
JP4386367B2 (en) 2006-02-08 2009-12-16 株式会社国際電気通信基礎技術研究所 Communication robot improvement system
EP1901153A1 (en) 2006-09-12 2008-03-19 OFFIS e.V. Control system for unmanned 4-rotor-helicopter
JP2009093308A (en) * 2007-10-05 2009-04-30 Hitachi Industrial Equipment Systems Co Ltd Robot system
JP5105596B2 (en) * 2007-10-30 2012-12-26 株式会社Ihi Travel route determination map creation device and travel route determination map creation method for autonomous mobile body
JP2009136987A (en) * 2007-12-10 2009-06-25 Toyota Motor Corp Mobile robot and method of correcting floor surface shape data
JP5349804B2 (en) * 2008-01-10 2013-11-20 株式会社日立産機システム Mobile robot system and control method thereof
US8060270B2 (en) 2008-02-29 2011-11-15 The Boeing Company System and method for inspection of structures and objects by swarm of remote unmanned vehicles
JP2009294713A (en) 2008-06-02 2009-12-17 Sanyo Electric Co Ltd Checking system, controller, checking method, and control program
JP5141507B2 (en) 2008-08-25 2013-02-13 村田機械株式会社 Autonomous mobile device
FR2938774A1 (en) * 2008-11-27 2010-05-28 Parrot DEVICE FOR CONTROLLING A DRONE
CN103025609A (en) * 2010-05-26 2013-04-03 威罗门飞行公司 Reconfigurable battery-operated vehicle system
US20140061376A1 (en) 2010-05-26 2014-03-06 Aerovironment Inc Reconfigurable battery-operated vehicle system
JP5614844B2 (en) * 2011-04-26 2014-10-29 学校法人千葉工業大学 Multi-rotor helicopter cross wind stabilizer and multi-rotor helicopter equipped with the same
CN102435188B (en) * 2011-09-15 2013-10-02 南京航空航天大学 Monocular vision/inertia autonomous navigation method for indoor environment
CN202600150U (en) 2012-05-17 2012-12-12 北京必威易激光科技有限公司 Intelligent low-altitude remote sensing surveying and mapping system
CN103455036B (en) 2012-06-05 2018-04-27 国家电网公司 A kind of scene aerial patrol method and aircraft
JP6029446B2 (en) * 2012-12-13 2016-11-24 セコム株式会社 Autonomous flying robot
CN103144770B (en) * 2013-03-19 2015-10-28 南京航空航天大学 A kind of entirely keep away barrier navigation minute vehicle from master control environment of entering the room
CN103365295B (en) 2013-06-29 2015-09-30 天津大学 Based on the autonomous hover control system of four rotor unmanned aircrafts and the method for DSP
CN203342367U (en) 2013-07-01 2013-12-18 昊翔电能运动科技(昆山)有限公司 Multiaxial flying machine
CN103365298B (en) * 2013-07-05 2017-06-20 深圳市大疆创新科技有限公司 The flight assisting system and method for unmanned vehicle
CN103697889B (en) 2013-12-29 2016-05-25 北京航空航天大学 A kind of unmanned plane independent navigation and localization method based on multi-model Distributed filtering
WO2015107623A1 (en) * 2014-01-15 2015-07-23 パイオニア株式会社 Management system and position specification method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
None *

Also Published As

Publication number Publication date
WO2016002236A1 (en) 2016-01-07
US10359778B2 (en) 2019-07-23
TWI578132B (en) 2017-04-11
JP2016015628A (en) 2016-01-28
TW201602748A (en) 2016-01-16
US20170139410A1 (en) 2017-05-18
EP3147743A4 (en) 2017-05-31
CN106662880B (en) 2019-08-06
EP3147743A1 (en) 2017-03-29
JP6486024B2 (en) 2019-03-20
CN106662880A (en) 2017-05-10

Similar Documents

Publication Publication Date Title
EP3147743B1 (en) Indoor monitoring system and method for structure
CN109911188B (en) Bridge detection unmanned aerial vehicle system in non-satellite navigation and positioning environment
JP6375503B2 (en) Flight type inspection apparatus and inspection method
EP3460392A2 (en) Positioning system for aerial non-destructive inspection
DK2702382T3 (en) METHOD AND SYSTEM FOR INSPECTION OF A SURFACE ERROR FOR MATERIAL ERROR
EP3474109B1 (en) Measuring device, control device for unmanned aerial vehicle and method for controlling unmanned aerial vehicle
EP2472221B1 (en) Flight Control System for Flying Object
CN111943043A (en) Guide information display device, crane provided with same, and guide information display method
JP2016111414A (en) Flying body position detection system and flying body
EP4063985A1 (en) Aerial inspection system
JP2020072366A (en) Imaging system and imaging method
US20220099442A1 (en) Surveying System
US11913789B2 (en) Inspection management device, inspection management method, and recording medium to store program
KR101970789B1 (en) Apparatus for detecting a pipe position and pose, and its method for detecting a pipe position and pose
JP2008057243A (en) Method and device for managing building of bridge
CN102285590B (en) Container lifting tool gripping operation sensing system and method based on infrared structure light
US9334031B2 (en) System for attitude control of underwater vehicle and method for attitude control of underwater vehicle
JP2020191026A (en) Patrol inspection system
US20240045449A1 (en) Method for controlling a drone along a shaft
JP2020072465A (en) Imaging system and imaging method
JP2020070011A (en) Imaging system and imaging method
JP6737521B2 (en) Height position acquisition system and imaging system, height position acquisition method, and imaging method
KR102504743B1 (en) Position correction device and correction method of inspection drone based on the model of the facility
JP6715542B1 (en) Aircraft, inspection method and inspection system
JP2023070120A (en) Autonomous flight control method, autonomous flight control apparatus and autonomous flight control system

Legal Events

Date Code Title Description
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE

PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE

17P Request for examination filed

Effective date: 20161223

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

AX Request for extension of the european patent

Extension state: BA ME

A4 Supplementary search report drawn up and despatched

Effective date: 20170503

RIC1 Information provided on ipc code assigned before grant

Ipc: G01C 15/00 20060101ALI20170425BHEP

Ipc: G05D 1/10 20060101AFI20170425BHEP

Ipc: G01B 21/00 20060101ALI20170425BHEP

Ipc: G01N 21/84 20060101ALI20170425BHEP

Ipc: G05D 1/00 20060101ALI20170425BHEP

DAX Request for extension of the european patent (deleted)
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20180122

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

INTG Intention to grant announced

Effective date: 20180926

GRAS Grant fee paid

Free format text: ORIGINAL CODE: EPIDOSNIGR3

GRAJ Information related to disapproval of communication of intention to grant by the applicant or resumption of examination proceedings by the epo deleted

Free format text: ORIGINAL CODE: EPIDOSDIGR1

GRAL Information related to payment of fee for publishing/printing deleted

Free format text: ORIGINAL CODE: EPIDOSDIGR3

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

GRAR Information related to intention to grant a patent recorded

Free format text: ORIGINAL CODE: EPIDOSNIGR71

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

GRAA (expected) grant

Free format text: ORIGINAL CODE: 0009210

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE PATENT HAS BEEN GRANTED

INTC Intention to grant announced (deleted)
INTG Intention to grant announced

Effective date: 20190220

AK Designated contracting states

Kind code of ref document: B1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

REG Reference to a national code

Ref country code: GB

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: CH

Ref legal event code: EP

Ref country code: AT

Ref legal event code: REF

Ref document number: 1116515

Country of ref document: AT

Kind code of ref document: T

Effective date: 20190415

REG Reference to a national code

Ref country code: DE

Ref legal event code: R096

Ref document number: 602015027737

Country of ref document: DE

REG Reference to a national code

Ref country code: IE

Ref legal event code: FG4D

REG Reference to a national code

Ref country code: NL

Ref legal event code: MP

Effective date: 20190403

REG Reference to a national code

Ref country code: LT

Ref legal event code: MG4D

REG Reference to a national code

Ref country code: AT

Ref legal event code: MK05

Ref document number: 1116515

Country of ref document: AT

Kind code of ref document: T

Effective date: 20190403

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: NL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: AL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: PT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190803

Ref country code: SE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: FI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: NO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190703

Ref country code: HR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: ES

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: LT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: CZ

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LV

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: RS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: BG

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190703

Ref country code: PL

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: GR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190704

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IS

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190803

Ref country code: AT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

REG Reference to a national code

Ref country code: DE

Ref legal event code: R097

Ref document number: 602015027737

Country of ref document: DE

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: DK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: EE

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: SK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: RO

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

PLBE No opposition filed within time limit

Free format text: ORIGINAL CODE: 0009261

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: SM

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

26N No opposition filed

Effective date: 20200106

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: TR

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: SI

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MC

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

REG Reference to a national code

Ref country code: CH

Ref legal event code: PL

REG Reference to a national code

Ref country code: BE

Ref legal event code: MM

Effective date: 20200131

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: LU

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20200120

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: BE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20200131

Ref country code: LI

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20200131

Ref country code: CH

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20200131

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: IE

Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

Effective date: 20200120

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MT

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

Ref country code: CY

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

Ref country code: MK

Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT

Effective date: 20190403

REG Reference to a national code

Ref country code: DE

Ref legal event code: R079

Ref document number: 602015027737

Country of ref document: DE

Free format text: PREVIOUS MAIN CLASS: G05D0001100000

Ipc: G05D0001460000

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: GB

Payment date: 20231130

Year of fee payment: 10

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: FR

Payment date: 20231212

Year of fee payment: 10

PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

Ref country code: DE

Payment date: 20231128

Year of fee payment: 10