US20230205211A1 - Protection of ultraviolet (uv) light source on mobile device - Google Patents
Protection of ultraviolet (uv) light source on mobile device Download PDFInfo
- Publication number
- US20230205211A1 US20230205211A1 US18/111,200 US202318111200A US2023205211A1 US 20230205211 A1 US20230205211 A1 US 20230205211A1 US 202318111200 A US202318111200 A US 202318111200A US 2023205211 A1 US2023205211 A1 US 2023205211A1
- Authority
- US
- United States
- Prior art keywords
- mobile robot
- sensor
- light source
- controller
- autonomous mobile
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000001514 detection method Methods 0.000 claims description 11
- 230000033001 locomotion Effects 0.000 claims description 9
- 230000008859 change Effects 0.000 claims description 7
- 244000005700 microbiome Species 0.000 claims description 3
- 241000894006 Bacteria Species 0.000 claims description 2
- 208000035473 Communicable disease Diseases 0.000 claims description 2
- 241000700605 Viruses Species 0.000 claims description 2
- 238000004659 sterilization and disinfection Methods 0.000 claims 1
- 238000004590 computer program Methods 0.000 description 6
- 238000000034 method Methods 0.000 description 5
- 230000003287 optical effect Effects 0.000 description 4
- 230000001681 protective effect Effects 0.000 description 3
- 239000007787 solid Substances 0.000 description 3
- 238000004140 cleaning Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000001627 detrimental effect Effects 0.000 description 1
- 238000009429 electrical wiring Methods 0.000 description 1
- 230000005670 electromagnetic radiation Effects 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0212—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
- G05D1/0214—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory in accordance with safety or protection criteria, e.g. avoiding hazardous areas
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0212—Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61L—METHODS OR APPARATUS FOR STERILISING MATERIALS OR OBJECTS IN GENERAL; DISINFECTION, STERILISATION OR DEODORISATION OF AIR; CHEMICAL ASPECTS OF BANDAGES, DRESSINGS, ABSORBENT PADS OR SURGICAL ARTICLES; MATERIALS FOR BANDAGES, DRESSINGS, ABSORBENT PADS OR SURGICAL ARTICLES
- A61L2/00—Methods or apparatus for disinfecting or sterilising materials or objects other than foodstuffs or contact lenses; Accessories therefor
- A61L2/02—Methods or apparatus for disinfecting or sterilising materials or objects other than foodstuffs or contact lenses; Accessories therefor using physical phenomena
- A61L2/08—Radiation
- A61L2/10—Ultraviolet radiation
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/0011—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots associated with a remote control arrangement
- G05D1/0016—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots associated with a remote control arrangement characterised by the operator's input device
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0231—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
- G05D1/0238—Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using obstacle or wall sensors
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/02—Control of position or course in two dimensions
- G05D1/021—Control of position or course in two dimensions specially adapted to land vehicles
- G05D1/0257—Control of position or course in two dimensions specially adapted to land vehicles using a radar
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61L—METHODS OR APPARATUS FOR STERILISING MATERIALS OR OBJECTS IN GENERAL; DISINFECTION, STERILISATION OR DEODORISATION OF AIR; CHEMICAL ASPECTS OF BANDAGES, DRESSINGS, ABSORBENT PADS OR SURGICAL ARTICLES; MATERIALS FOR BANDAGES, DRESSINGS, ABSORBENT PADS OR SURGICAL ARTICLES
- A61L2202/00—Aspects relating to methods or apparatus for disinfecting or sterilising materials or objects
- A61L2202/10—Apparatus features
- A61L2202/11—Apparatus for generating biocidal substances, e.g. vaporisers, UV lamps
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61L—METHODS OR APPARATUS FOR STERILISING MATERIALS OR OBJECTS IN GENERAL; DISINFECTION, STERILISATION OR DEODORISATION OF AIR; CHEMICAL ASPECTS OF BANDAGES, DRESSINGS, ABSORBENT PADS OR SURGICAL ARTICLES; MATERIALS FOR BANDAGES, DRESSINGS, ABSORBENT PADS OR SURGICAL ARTICLES
- A61L2202/00—Aspects relating to methods or apparatus for disinfecting or sterilising materials or objects
- A61L2202/10—Apparatus features
- A61L2202/14—Means for controlling sterilisation processes, data processing, presentation and storage means, e.g. sensors, controllers, programs
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61L—METHODS OR APPARATUS FOR STERILISING MATERIALS OR OBJECTS IN GENERAL; DISINFECTION, STERILISATION OR DEODORISATION OF AIR; CHEMICAL ASPECTS OF BANDAGES, DRESSINGS, ABSORBENT PADS OR SURGICAL ARTICLES; MATERIALS FOR BANDAGES, DRESSINGS, ABSORBENT PADS OR SURGICAL ARTICLES
- A61L2202/00—Aspects relating to methods or apparatus for disinfecting or sterilising materials or objects
- A61L2202/10—Apparatus features
- A61L2202/16—Mobile applications, e.g. portable devices, trailers, devices mounted on vehicles
-
- G05D2201/02—
Definitions
- Mobile devices such as mobile robots
- Some mobile robots can detect humans in a determined path of direction, and change the determined path.
- a device may be a mobile robot that may include a motor to drive a drive system to move the mobile robot in an area, and a light source to output ultraviolet light.
- the device may include at least one first sensor to determine at least one of an orientation of the mobile robot, a location of the mobile robot, and/or when the light source is within a predetermined distance of an object in the area.
- the device may include a controller, communicatively coupled to the drive system, the light source, and the at least one first sensor to control the drive system so as to stop or move the mobile robot before the light source is within the predetermined distance of the object based on at least a signal received from the at least one first sensor.
- FIGS. 1 - 3 show a plurality of external views of a mobile robot having sensors to protect an ultraviolet (UV) light source according to implementations of the disclosed subject matter.
- UV ultraviolet
- FIG. 4 shows an example configuration of the mobile robot of FIGS. 1 - 3 according to an implementation of the disclosed subject matter.
- FIG. 5 shows a network configuration which may include a plurality of mobile robots according to implementations of the disclosed subject matter.
- a mobile robot may emit ultraviolet (UV) light from a UV light source.
- the UV light output by the light source may be used to disinfect an object, at least a portion of a room, a predetermined area, or the like.
- the mobile robot may be used as part of a regular cleaning cycle of a room, building, or the like, and may prevent and/or reduce the spread of infectious diseases, viruses, bacteria, and other types of harmful organic microorganisms in the environment by breaking down their DNA-structure with UV light.
- the mobile robot may reduce human error in cleaning an area, room, building, or the like by tracking the location and/or intensity (e.g., optical power of UV light) of light radiated, and determine which areas may need to be radiated and/or cleaned.
- intensity e.g., optical power of UV light
- the mobile robot may be operated manually, autonomously, and/or may receive control signals to control the movement of the mobile robot with a room, building, area, or the like when operating in a tele-operation mode.
- the UV light source of the mobile robot may be protected by one or more sensors disposed on the mobile robot.
- the sensors may prevent the mobile robot from moving to a location having one or more objects that may damage the UV light source.
- the one or more sensors may be communicatively coupled to a controller, which may control the movement of the mobile robot based on signals received from the one or more sensors. This differs from traditional mobile robots, which may avoid objects and/or humans to prevent damaging the objects or harming humans. That is, implementations of the disclosed subject matter provide sensors to detect obstacles which may potentially damage the mobile robot, such as its UV light source.
- a protective shield may be used to at least partially cover and/or enclose the UV light source, the magnitude of the UV light (i.e., the optical power of the UV light) may physically deform the protective shield over time, and/or the protective shield may attenuate the amount of UV light that may radiate a predetermined area, which may be detrimental in preventing the spread of harmful organic microorganisms.
- FIGS. 1 - 3 show a plurality of external views of a mobile robot 100 with sensors to protect an ultraviolet (UV) light source according to implementations of the disclosed subject matter.
- the mobile robot 100 may include at least a first sensor 102 (shown as sensor 102 a and 102 b in FIG. 2 ), a light source 104 to output ultraviolet light, at least a second sensor 106 , a drive system 108 , a user interface 110 , and/or a stop button 112 .
- a controller e.g., controller 114 shown in FIG.
- the at least one first sensor 102 may be communicatively coupled to the at least one first sensor 102 , the light source 104 , the at least one second sensor 106 , the drive system 108 , the user interface 110 and the stop button 112 , may control the operations of the mobile robot 100 .
- the at least one first sensor 102 may determine at least one of an orientation of the mobile robot 100 (e.g., a direction that a front side and/or a first side of a robot is facing), a location of the mobile robot 100 (e.g., a location of the mobile robot 100 in an area), and/or when the light source 104 is within a predetermined distance of an object 113 (which may include 113 a , 113 b ) in the area.
- the at least one first sensor 102 may have a field of view of 70 degrees diagonally.
- the at least one sensor 102 may have a detection distance of 0.2-4 meters. As shown in FIGS. 1 - 3 , the at least one first sensor 102 may be disposed over the light source 104 .
- the at least one first sensor 102 may include a first side sensor disposed on a first side of the mobile robot 100 and a second side sensor that may be disposed on a second side of the device.
- sensor 102 a may be disposed on a first side (e.g., a front side) of the mobile robot 100
- sensor 102 b may be disposed on a second side (e.g., a back side) of the mobile robot 100
- sensors on two sides of the robot are shown in FIG. 2 , there may be a plurality of sensors disposed on different sides of the mobile robot 102 to prevent an object from damaging the light source 104 .
- sensor 102 a and/or sensor 102 b may be disposed over the light source 104 .
- the light source 104 may be one or more bulbs, one or more lamps, and/or an array of light emitting diodes (LEDs) or organic light emitting diodes (OLEDs) to emit UV light (e.g., light having a wavelength of 10 nm-400 nm).
- the intensity i.e., optical power output
- the controller 114 may also turn on or off a portion or all of the devices (e.g., bulbs, lamps, LEDs, OLEDs) of the light source 104 .
- the at least one second sensor 106 may be communicatively coupled to the controller 114 shown in FIG. 4 , and the controller 114 may control the drive system 108 so as to stop and/or move the mobile robot 100 before the light source 104 is within the predetermined distance of an object 113 (e.g., at least a portion of object 113 a , 113 b ) based on at least a signal received from the at least one first sensor 102 and/or a signal received from the at least one second sensor 106 .
- the light source 104 may be disposed over the at least one second sensor 106 .
- the at least one second sensor 106 may be oriented in a direction towards the light source 104 .
- the at least one first sensor 102 may be arranged so as to have a detection direction that is about parallel to a surface over which the mobile robot is travelling (e.g., the floor, ground, or the like) and/or towards the surface (e.g., a downward detection direction), and the at least one second sensor 106 may have a detection direction oriented upwards from the surface (e.g., in an upwards direction) towards the light source 104 .
- the at least one second sensor 106 may be a time-of-flight sensor, an ultrasonic sensor, a two-dimensional Light Detection and Ranging (LiDAR) sensor, a three-dimensional LiDAR sensor, and/or a radar (radio detection and ranging) sensor, or the like.
- the at least one second sensor 106 may have a field of view of 20-27 degrees. In some implementations, the at least one second sensor 106 may have a detection distance of 0.05-4 meters.
- the mobile robot 100 may include a motor to drive the drive system 108 to move the mobile robot in an area, such as a room, a building, or the like.
- the drive system 108 may include wheels, which may be adjustable so that the drive system 108 may control the direction of the mobile robot 100 .
- the mobile robot 100 may include a base with the drive system 108 , and at least one first sensor 102 may be disposed on the base.
- the at least one first sensor 102 may be a two-dimensional Light Detection and Ranging (LiDAR) sensor, a three-dimensional LiDAR sensor, three-dimensional cameras, or the like.
- LiDAR Light Detection and Ranging
- the controller 114 may control and/or operate the mobile robot 100 in an operation mode which may be a manual mode, an autonomous mode, and/or a tele-operation mode.
- the controller 114 may receive on or more control signals from the user interface 110 and/or the stop button 112 .
- a user may control the movement, direction, and/or stop the motion of the mobile robot 100 by making one or more selections on the user interface 110 .
- the stop button 112 may be an emergency stop (ESTOP) button which may stop all operations and/or movement of the mobile robot 100 when selected.
- the controller 114 may receive at least one control signal via a network interface 116 (shown in FIG. 4 ) when operating when operating in the tele-operation mode.
- the network interface may receive control signals via network 130 from server 140 , database 150 , and/or remote platform 160 , as described below in connection with FIG. 5 .
- the at least one first sensor 102 may be a time-of-flight sensor, a stereo vision sensor, a two-dimensional LiDAR sensor, a three-dimensional LiDAR sensor, or the like.
- the at least one first sensor 102 may detect a geometry of one or more objects (e.g., object 113 , having portions 113 a and/or 113 b ) that may be in the path of the mobile robot 100 .
- the output of the at least one first sensor 102 may be, for example, a point cloud of the one or more objects in the path of the mobile robot 100 .
- the controller 114 may reduce an allowed maximum velocity of the mobile robot 100 for safety, so as to reduce and/or eliminate having the light source 104 impact the one or more objects which may damage the light source 104 .
- the controller 114 may determine whether the desired path (i.e., a planned path) may cause the mobile robot to impact the one or more objects (e.g., object 113 ).
- the controller 114 may receive a user command via the user interface 110 , the stop button 112 , and/or via the network interface 116 . If the desired path may end with a collision, or may place the mobile robot 100 within a predetermined distance of an object which may cause a collision, the controller 114 may change the path of motion, or stop movement, of the mobile robot 100 by controlling the drive system 108 .
- the at least one first sensor 102 is a stereo vision sensor
- images from two sensors i.e., where the two sensors may be part of the stereo vision sensor of the at least one first sensor 102
- the global shutter may be configured so that the two sensors of the stereo vision sensor may capture images about simultaneously.
- One or more features may be determined from the captured images, and be compared to one another to determine portions that are matching.
- the focal length of the two sensors of the stereo vision sensor and the distance between the two sensors (e.g., about 6 cm) may be stored in memory 118 and/or fixed storage 120 (shown in FIG.
- the controller 114 and/or the at least one first sensor 102 may use the captured images and the stored values to determine the distance from the at least one first sensor 102 to the object 113 .
- the at least one sensor 102 may include at least one laser, LED, and/or OLED, to radiate one or more points on surfaces of objects, when the objects may be without identifying features (e.g., blank walls).
- protection of the light source 104 may be increased by using the at least one second sensor 106 which may be disposed below the light source 104 .
- the at least one second sensor 106 may be configured so that a detection path is towards the light source 104 (e.g., an upward path that may be at an angle from the surface over which the mobile robot 100 may travel).
- the at least one second sensor 106 may be disposed above the light source 104 .
- the at least one second sensor 106 may include one sensor that is disposed above the light source 104 , and another sensor that is disposed below the light source 104 .
- the at least one second sensor 106 may be a time-of-flight (TOF) sensor. At least one photon of light may be output by the at least one second sensor 106 , and may be transmitted through the air. When the at least one photon of light radiates the object 113 , a portion of the light may be reflected by the object 113 may return to a receiver portion of the at least one second sensor 106 .
- the at least one second sensor 106 may calculate the time between sending the at least one photon of light and receiving the reflection, and multiply this value by the speed of light in air, to determine the distance between the at least one second sensor 106 and the object 113 . Typically, the calculated distance may be larger than the distance between the object 113 and the light source 104 .
- the distance may differ from place to place, for example, because of different ceiling heights (e.g., which may reflect the one or more photons of light differently).
- the controller 114 may stop the drive system 108 and/or change the direction of the mobile robot 100 . This determination by the controller 114 may be based on the present direction of the mobile robot 100 , the orientation of the mobile robot 100 , the location of the at least one second sensor 106 , and/or the physical dimension of the mobile robot 100 .
- the at least one second sensor 106 may stop the mobile robot 100 by transmitting a signal to the controller 114 to change the path direction of the mobile robot 100 when operating in the autonomous mode.
- the at least one second sensor 106 may provide a control signal to the drive system 108 to stop the movement of the mobile robot 100 .
- FIG. 4 shows example components of the mobile robot 100 suitable for providing the implementations of the disclosed subject matter.
- the mobile robot 100 may include a bus 122 which interconnects major components of the mobile robot 100 , such as the drive system 108 , a network interface 116 operable to communicate with one or more remote devices via a suitable network connection, the controller 114 , a memory 118 such as Random Access Memory (RAM), Read Only Memory (ROM), flash RAM, or the like, the stop button 112 , the light source 104 , the at least one first sensor 102 , a user interface 110 that may include one or more controllers and associated user input devices such as a keyboard, touch screen, and the like, a fixed storage 120 such as a hard drive, flash storage, and the like, and the at least one second sensor 106 .
- RAM Random Access Memory
- ROM Read Only Memory
- the bus 122 allows data communication between the controller 114 and one or more memory components, which may include RAM, ROM, and other memory, as previously noted.
- RAM is the main memory into which an operating system and application programs are loaded.
- a ROM or flash memory component can contain, among other code, the Basic Input-Output system (BIOS) which controls basic hardware operation such as the interaction with peripheral components.
- BIOS Basic Input-Output system
- Applications resident with the mobile robot 100 are generally stored on and accessed via a computer readable medium (e.g., fixed storage 120 ), such as a solid state drive, hard disk drive, an optical drive, solid state drive, or other storage medium.
- the network interface 116 may provide a direct connection to a remote server (e.g., server 140 , database 150 , and/or remote platform 160 shown in FIG. 5 ) via a wired or wireless connection (e.g., network 130 shown in FIG. 5 ).
- the network interface 116 may provide such connection using any suitable technique and protocol as will be readily understood by one of skill in the art, including digital cellular telephone, WiFi, Bluetooth®, near-field, and the like.
- the network interface 116 may allow the mobile robot 100 to communicate with other computers via one or more local, wide-area, or other communication networks, as described in further detail below.
- the mobile robot may transmit data via the network interface to the remote server that may include a path of operation, the surfaces and/or areas radiated with UV light, and the like.
- Code to implement the present disclosure can be stored in computer-readable storage media such as one or more of the memory 118 , fixed storage 120 , or on a remote storage location.
- FIG. 5 shows an example network arrangement according to an implementation of the disclosed subject matter.
- Mobile robot 100 described above, and/or a similar mobile robot 200 may connect to other devices via network 130 .
- the network 130 may be a local network, wide-area network, the Internet, or any other suitable communication network or networks, and may be implemented on any suitable platform including wired and/or wireless networks.
- the mobile robot 100 and/or mobile robot 200 may communicate with one another, and/or may communicate with one or more remote devices, such as server 140 , database 150 , and/or remote platform 160 .
- the remote devices may be directly accessible by the mobile robot 100 , 200 or one or more other devices may provide intermediary access such as where a server 140 provides access to resources stored in a database 150 .
- the mobile robot 100 , 200 may access remote platform 160 or services provided by remote platform 160 such as cloud computing arrangements and services.
- the remote platform 160 may include one or more servers 140 and/or databases 150 .
- implementations of the presently disclosed subject matter may include or be embodied in the form of computer-implemented processes and apparatuses for practicing those processes.
- Implementations also may be embodied in the form of a computer program product having computer program code containing instructions embodied in non-transitory and/or tangible media, such as solid state drives, DVDs, CD-ROMs, hard drives, USB (universal serial bus) drives, or any other machine readable storage medium, such that when the computer program code is loaded into and executed by a computer, the computer becomes an apparatus for practicing implementations of the disclosed subject matter.
- Implementations also may be embodied in the form of computer program code, for example, whether stored in a storage medium, loaded into and/or executed by a computer, or transmitted over some transmission medium, such as over electrical wiring or cabling, through fiber optics, or via electromagnetic radiation, such that when the computer program code is loaded into and executed by a computer, the computer becomes an apparatus for practicing implementations of the disclosed subject matter.
- the computer program code segments configure the microprocessor to create specific logic circuits.
- a set of computer-readable instructions stored on a computer-readable storage medium may be implemented by a general-purpose processor, which may transform the general-purpose processor or a device containing the general-purpose processor into a special-purpose device configured to implement or carry out the instructions. Implementations may include using hardware that has a processor, such as a general purpose microprocessor and/or an Application Specific Integrated Circuit (ASIC) that embodies all or part of the techniques according to implementations of the disclosed subject matter in hardware and/or firmware.
- the processor may be coupled to memory, such as RAM, ROM, flash memory, a hard disk or any other device capable of storing electronic information.
- the memory may store instructions adapted to be executed by the processor to perform the techniques according to implementations of the disclosed subject matter.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Aviation & Aerospace Engineering (AREA)
- Automation & Control Theory (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Public Health (AREA)
- Mathematical Physics (AREA)
- General Health & Medical Sciences (AREA)
- Animal Behavior & Ethology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Epidemiology (AREA)
- Computing Systems (AREA)
- Veterinary Medicine (AREA)
- Theoretical Computer Science (AREA)
- Electromagnetism (AREA)
- Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
- Manipulator (AREA)
- Pens And Brushes (AREA)
- Heating, Cooling, Or Curing Plastics Or The Like In General (AREA)
- Orthopedics, Nursing, And Contraception (AREA)
Abstract
Implementations of the disclosed subject matter provide a device of a mobile robot may include a motor to drive a drive system to move the mobile robot in an area, and a light source to output ultraviolet light. The device may include at least one first sensor to determine at least one of an orientation of the mobile robot, a location of the mobile robot, and/or when the light source is within a predetermined distance of an object in the area. The device may include a controller, communicatively coupled to the drive system, the light source, and the at least one first sensor to control the drive system so as to stop or move the mobile robot before the light source is within the predetermined distance of the object based on at least a signal received from the at least one first sensor.
Description
- Mobile devices, such as mobile robots, can be operated so as to change direction after striking a surface, such as a wall. Some mobile robots can detect humans in a determined path of direction, and change the determined path.
- According to an implementation of the disclosed subject matter, a device may be a mobile robot that may include a motor to drive a drive system to move the mobile robot in an area, and a light source to output ultraviolet light. The device may include at least one first sensor to determine at least one of an orientation of the mobile robot, a location of the mobile robot, and/or when the light source is within a predetermined distance of an object in the area. The device may include a controller, communicatively coupled to the drive system, the light source, and the at least one first sensor to control the drive system so as to stop or move the mobile robot before the light source is within the predetermined distance of the object based on at least a signal received from the at least one first sensor.
- Additional features, advantages, and implementations of the disclosed subject matter may be set forth or apparent from consideration of the following detailed description, drawings, and claims. Moreover, it is to be understood that both the foregoing summary and the following detailed description are illustrative and are intended to provide further explanation without limiting the scope of the claims.
- The accompanying drawings, which are included to provide a further understanding of the disclosed subject matter, are incorporated in and constitute a part of this specification. The drawings also illustrate implementations of the disclosed subject matter and together with the detailed description serve to explain the principles of implementations of the disclosed subject matter. No attempt is made to show structural details in more detail than may be necessary for a fundamental understanding of the disclosed subject matter and various ways in which it may be practiced.
-
FIGS. 1-3 show a plurality of external views of a mobile robot having sensors to protect an ultraviolet (UV) light source according to implementations of the disclosed subject matter. -
FIG. 4 shows an example configuration of the mobile robot ofFIGS. 1-3 according to an implementation of the disclosed subject matter. -
FIG. 5 shows a network configuration which may include a plurality of mobile robots according to implementations of the disclosed subject matter. - In implementations of the disclosed subject matter, a mobile robot may emit ultraviolet (UV) light from a UV light source. The UV light output by the light source may be used to disinfect an object, at least a portion of a room, a predetermined area, or the like. The mobile robot may be used as part of a regular cleaning cycle of a room, building, or the like, and may prevent and/or reduce the spread of infectious diseases, viruses, bacteria, and other types of harmful organic microorganisms in the environment by breaking down their DNA-structure with UV light. The mobile robot may reduce human error in cleaning an area, room, building, or the like by tracking the location and/or intensity (e.g., optical power of UV light) of light radiated, and determine which areas may need to be radiated and/or cleaned.
- The mobile robot may be operated manually, autonomously, and/or may receive control signals to control the movement of the mobile robot with a room, building, area, or the like when operating in a tele-operation mode.
- The UV light source of the mobile robot may be protected by one or more sensors disposed on the mobile robot. The sensors may prevent the mobile robot from moving to a location having one or more objects that may damage the UV light source. The one or more sensors may be communicatively coupled to a controller, which may control the movement of the mobile robot based on signals received from the one or more sensors. This differs from traditional mobile robots, which may avoid objects and/or humans to prevent damaging the objects or harming humans. That is, implementations of the disclosed subject matter provide sensors to detect obstacles which may potentially damage the mobile robot, such as its UV light source.
- Although a protective shield may be used to at least partially cover and/or enclose the UV light source, the magnitude of the UV light (i.e., the optical power of the UV light) may physically deform the protective shield over time, and/or the protective shield may attenuate the amount of UV light that may radiate a predetermined area, which may be detrimental in preventing the spread of harmful organic microorganisms.
-
FIGS. 1-3 show a plurality of external views of amobile robot 100 with sensors to protect an ultraviolet (UV) light source according to implementations of the disclosed subject matter. Themobile robot 100 may include at least a first sensor 102 (shown assensor FIG. 2 ), alight source 104 to output ultraviolet light, at least asecond sensor 106, adrive system 108, auser interface 110, and/or astop button 112. A controller (e.g.,controller 114 shown inFIG. 4 and described below) may be communicatively coupled to the at least onefirst sensor 102, thelight source 104, the at least onesecond sensor 106, thedrive system 108, theuser interface 110 and thestop button 112, may control the operations of themobile robot 100. - The at least one first sensor 102 (including
sensors FIG. 2 ) may determine at least one of an orientation of the mobile robot 100 (e.g., a direction that a front side and/or a first side of a robot is facing), a location of the mobile robot 100 (e.g., a location of themobile robot 100 in an area), and/or when thelight source 104 is within a predetermined distance of an object 113 (which may include 113 a, 113 b) in the area. In some implementations, the at least onefirst sensor 102 may have a field of view of 70 degrees diagonally. The at least onesensor 102 may have a detection distance of 0.2-4 meters. As shown inFIGS. 1-3 , the at least onefirst sensor 102 may be disposed over thelight source 104. - The at least one
first sensor 102 may include a first side sensor disposed on a first side of themobile robot 100 and a second side sensor that may be disposed on a second side of the device. For example, as shown inFIG. 2 ,sensor 102 a may be disposed on a first side (e.g., a front side) of themobile robot 100, andsensor 102 b may be disposed on a second side (e.g., a back side) of themobile robot 100. Although sensors on two sides of the robot are shown inFIG. 2 , there may be a plurality of sensors disposed on different sides of themobile robot 102 to prevent an object from damaging thelight source 104. In some implementations,sensor 102 a and/orsensor 102 b may be disposed over thelight source 104. - The
light source 104 may be one or more bulbs, one or more lamps, and/or an array of light emitting diodes (LEDs) or organic light emitting diodes (OLEDs) to emit UV light (e.g., light having a wavelength of 10 nm-400 nm). The intensity (i.e., optical power output) may be controlled by thecontroller 114, which may also turn on or off a portion or all of the devices (e.g., bulbs, lamps, LEDs, OLEDs) of thelight source 104. - The at least one
second sensor 106 may be communicatively coupled to thecontroller 114 shown inFIG. 4 , and thecontroller 114 may control thedrive system 108 so as to stop and/or move themobile robot 100 before thelight source 104 is within the predetermined distance of an object 113 (e.g., at least a portion ofobject first sensor 102 and/or a signal received from the at least onesecond sensor 106. As shown inFIG. 1 , thelight source 104 may be disposed over the at least onesecond sensor 106. In some implementations, the at least onesecond sensor 106 may be oriented in a direction towards thelight source 104. For example, the at least onefirst sensor 102 may be arranged so as to have a detection direction that is about parallel to a surface over which the mobile robot is travelling (e.g., the floor, ground, or the like) and/or towards the surface (e.g., a downward detection direction), and the at least onesecond sensor 106 may have a detection direction oriented upwards from the surface (e.g., in an upwards direction) towards thelight source 104. - In some implementations, the at least one
second sensor 106 may be a time-of-flight sensor, an ultrasonic sensor, a two-dimensional Light Detection and Ranging (LiDAR) sensor, a three-dimensional LiDAR sensor, and/or a radar (radio detection and ranging) sensor, or the like. The at least onesecond sensor 106 may have a field of view of 20-27 degrees. In some implementations, the at least onesecond sensor 106 may have a detection distance of 0.05-4 meters. - The
mobile robot 100 may include a motor to drive thedrive system 108 to move the mobile robot in an area, such as a room, a building, or the like. Thedrive system 108 may include wheels, which may be adjustable so that thedrive system 108 may control the direction of themobile robot 100. - In some implementations, the
mobile robot 100 may include a base with thedrive system 108, and at least onefirst sensor 102 may be disposed on the base. In this implementation, the at least onefirst sensor 102 may be a two-dimensional Light Detection and Ranging (LiDAR) sensor, a three-dimensional LiDAR sensor, three-dimensional cameras, or the like. - The
controller 114 may control and/or operate themobile robot 100 in an operation mode which may be a manual mode, an autonomous mode, and/or a tele-operation mode. In the manual mode, thecontroller 114 may receive on or more control signals from theuser interface 110 and/or thestop button 112. For example, a user may control the movement, direction, and/or stop the motion of themobile robot 100 by making one or more selections on theuser interface 110. Thestop button 112 may be an emergency stop (ESTOP) button which may stop all operations and/or movement of themobile robot 100 when selected. In some implementations, thecontroller 114 may receive at least one control signal via a network interface 116 (shown inFIG. 4 ) when operating when operating in the tele-operation mode. For example, the network interface may receive control signals vianetwork 130 fromserver 140,database 150, and/orremote platform 160, as described below in connection withFIG. 5 . - In some implementations, the at least one
first sensor 102 may be a time-of-flight sensor, a stereo vision sensor, a two-dimensional LiDAR sensor, a three-dimensional LiDAR sensor, or the like. For example, while themobile robot 100 is moving in a direction, the at least onefirst sensor 102 may detect a geometry of one or more objects (e.g.,object 113, havingportions 113 a and/or 113 b) that may be in the path of themobile robot 100. The output of the at least onefirst sensor 102 may be, for example, a point cloud of the one or more objects in the path of themobile robot 100. In this example, there may be nothing (e.g., no objects) in the path of themobile robot 100 up to the detection range of the at least onefirst sensor 102, which may be about 4 meters. When the at least onefirst sensor 102 may determine that one or more objects (e.g., object 113) are in the path of themobile robot 100, thecontroller 114 may reduce an allowed maximum velocity of themobile robot 100 for safety, so as to reduce and/or eliminate having thelight source 104 impact the one or more objects which may damage thelight source 104. For example, when thecontroller 114 is operating themobile robot 100 in the autonomous mode, thecontroller 114 may determine whether the desired path (i.e., a planned path) may cause the mobile robot to impact the one or more objects (e.g., object 113). When thecontroller 114 is operating themobile robot 100 in the manual or tele-operated modes, thecontroller 114 may receive a user command via theuser interface 110, thestop button 112, and/or via thenetwork interface 116. If the desired path may end with a collision, or may place themobile robot 100 within a predetermined distance of an object which may cause a collision, thecontroller 114 may change the path of motion, or stop movement, of themobile robot 100 by controlling thedrive system 108. - When the at least one
first sensor 102 is a stereo vision sensor, images from two sensors (i.e., where the two sensors may be part of the stereo vision sensor of the at least one first sensor 102) within a known distance from one another distance may be captured at a predetermined point in time, and/or at predetermined time intervals with a global shutter. The global shutter may be configured so that the two sensors of the stereo vision sensor may capture images about simultaneously. One or more features may be determined from the captured images, and be compared to one another to determine portions that are matching. As the focal length of the two sensors of the stereo vision sensor and the distance between the two sensors (e.g., about 6 cm) may be stored inmemory 118 and/or fixed storage 120 (shown inFIG. 4 ), thecontroller 114 and/or the at least onefirst sensor 102 may use the captured images and the stored values to determine the distance from the at least onefirst sensor 102 to theobject 113. In some implementations, the at least onesensor 102 may include at least one laser, LED, and/or OLED, to radiate one or more points on surfaces of objects, when the objects may be without identifying features (e.g., blank walls). - In some implementations, protection of the
light source 104 may be increased by using the at least onesecond sensor 106 which may be disposed below thelight source 104. The at least onesecond sensor 106 may be configured so that a detection path is towards the light source 104 (e.g., an upward path that may be at an angle from the surface over which themobile robot 100 may travel). In some implementations, the at least onesecond sensor 106 may be disposed above thelight source 104. In another implementation, the at least onesecond sensor 106 may include one sensor that is disposed above thelight source 104, and another sensor that is disposed below thelight source 104. - For example, the at least one
second sensor 106 may be a time-of-flight (TOF) sensor. At least one photon of light may be output by the at least onesecond sensor 106, and may be transmitted through the air. When the at least one photon of light radiates theobject 113, a portion of the light may be reflected by theobject 113 may return to a receiver portion of the at least onesecond sensor 106. The at least onesecond sensor 106 may calculate the time between sending the at least one photon of light and receiving the reflection, and multiply this value by the speed of light in air, to determine the distance between the at least onesecond sensor 106 and theobject 113. Typically, the calculated distance may be larger than the distance between theobject 113 and thelight source 104. The distance may differ from place to place, for example, because of different ceiling heights (e.g., which may reflect the one or more photons of light differently). When the determined distance is less than the distance to the location of the at least one second sensor, thecontroller 114 may stop thedrive system 108 and/or change the direction of themobile robot 100. This determination by thecontroller 114 may be based on the present direction of themobile robot 100, the orientation of themobile robot 100, the location of the at least onesecond sensor 106, and/or the physical dimension of themobile robot 100. The at least onesecond sensor 106 may stop themobile robot 100 by transmitting a signal to thecontroller 114 to change the path direction of themobile robot 100 when operating in the autonomous mode. In another implementation, the at least onesecond sensor 106 may provide a control signal to thedrive system 108 to stop the movement of themobile robot 100. -
FIG. 4 shows example components of themobile robot 100 suitable for providing the implementations of the disclosed subject matter. Themobile robot 100 may include abus 122 which interconnects major components of themobile robot 100, such as thedrive system 108, anetwork interface 116 operable to communicate with one or more remote devices via a suitable network connection, thecontroller 114, amemory 118 such as Random Access Memory (RAM), Read Only Memory (ROM), flash RAM, or the like, thestop button 112, thelight source 104, the at least onefirst sensor 102, auser interface 110 that may include one or more controllers and associated user input devices such as a keyboard, touch screen, and the like, a fixedstorage 120 such as a hard drive, flash storage, and the like, and the at least onesecond sensor 106. - The
bus 122 allows data communication between thecontroller 114 and one or more memory components, which may include RAM, ROM, and other memory, as previously noted. Typically RAM is the main memory into which an operating system and application programs are loaded. A ROM or flash memory component can contain, among other code, the Basic Input-Output system (BIOS) which controls basic hardware operation such as the interaction with peripheral components. Applications resident with themobile robot 100 are generally stored on and accessed via a computer readable medium (e.g., fixed storage 120), such as a solid state drive, hard disk drive, an optical drive, solid state drive, or other storage medium. - The
network interface 116 may provide a direct connection to a remote server (e.g.,server 140,database 150, and/orremote platform 160 shown inFIG. 5 ) via a wired or wireless connection (e.g.,network 130 shown inFIG. 5 ). Thenetwork interface 116 may provide such connection using any suitable technique and protocol as will be readily understood by one of skill in the art, including digital cellular telephone, WiFi, Bluetooth®, near-field, and the like. For example, thenetwork interface 116 may allow themobile robot 100 to communicate with other computers via one or more local, wide-area, or other communication networks, as described in further detail below. The mobile robot may transmit data via the network interface to the remote server that may include a path of operation, the surfaces and/or areas radiated with UV light, and the like. - Many other devices or components (not shown) may be connected in a similar manner. Conversely, all of the components shown in
FIG. 4 need not be present to practice the present disclosure. The components can be interconnected in different ways from that shown. Code to implement the present disclosure can be stored in computer-readable storage media such as one or more of thememory 118, fixedstorage 120, or on a remote storage location. -
FIG. 5 shows an example network arrangement according to an implementation of the disclosed subject matter.Mobile robot 100 described above, and/or a similarmobile robot 200 may connect to other devices vianetwork 130. Thenetwork 130 may be a local network, wide-area network, the Internet, or any other suitable communication network or networks, and may be implemented on any suitable platform including wired and/or wireless networks. Themobile robot 100 and/ormobile robot 200 may communicate with one another, and/or may communicate with one or more remote devices, such asserver 140,database 150, and/orremote platform 160. The remote devices may be directly accessible by themobile robot server 140 provides access to resources stored in adatabase 150. Themobile robot remote platform 160 or services provided byremote platform 160 such as cloud computing arrangements and services. Theremote platform 160 may include one ormore servers 140 and/ordatabases 150. - More generally, various implementations of the presently disclosed subject matter may include or be embodied in the form of computer-implemented processes and apparatuses for practicing those processes. Implementations also may be embodied in the form of a computer program product having computer program code containing instructions embodied in non-transitory and/or tangible media, such as solid state drives, DVDs, CD-ROMs, hard drives, USB (universal serial bus) drives, or any other machine readable storage medium, such that when the computer program code is loaded into and executed by a computer, the computer becomes an apparatus for practicing implementations of the disclosed subject matter. Implementations also may be embodied in the form of computer program code, for example, whether stored in a storage medium, loaded into and/or executed by a computer, or transmitted over some transmission medium, such as over electrical wiring or cabling, through fiber optics, or via electromagnetic radiation, such that when the computer program code is loaded into and executed by a computer, the computer becomes an apparatus for practicing implementations of the disclosed subject matter. When implemented on a general-purpose microprocessor, the computer program code segments configure the microprocessor to create specific logic circuits.
- In some configurations, a set of computer-readable instructions stored on a computer-readable storage medium may be implemented by a general-purpose processor, which may transform the general-purpose processor or a device containing the general-purpose processor into a special-purpose device configured to implement or carry out the instructions. Implementations may include using hardware that has a processor, such as a general purpose microprocessor and/or an Application Specific Integrated Circuit (ASIC) that embodies all or part of the techniques according to implementations of the disclosed subject matter in hardware and/or firmware. The processor may be coupled to memory, such as RAM, ROM, flash memory, a hard disk or any other device capable of storing electronic information. The memory may store instructions adapted to be executed by the processor to perform the techniques according to implementations of the disclosed subject matter.
- The foregoing description, for purpose of explanation, has been described with reference to specific implementations. However, the illustrative discussions above are not intended to be exhaustive or to limit implementations of the disclosed subject matter to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The implementations were chosen and described in order to explain the principles of implementations of the disclosed subject matter and their practical applications, to thereby enable others skilled in the art to utilize those implementations as well as various implementations with various modifications as may be suited to the particular use contemplated.
Claims (11)
1. An autonomous mobile robot comprising:
a base that includes an autonomous drive system to move the autonomous mobile robot in an area;
a light source to output ultraviolet light that is disposed over the base;
at least one sensor to detect at least one selected from the group consisting of: an object in the area, and at least one surface in the area; and
a controller, communicatively coupled to the autonomous drive system, the light source, and the at least one sensor to control the autonomous drive system to perform at least one selected from the group consisting of: (1) control the autonomous drive system to move the autonomous mobile robot within an area, and (2) stop or change the path of the autonomous mobile robot before at least the light source impacts the object based on at least a signal received from the at least one sensor, wherein the controller is configured to controller is configured to control the light source to output ultraviolet light to disinfect one or more surfaces within the area, and wherein the controller is configured to track the one or more locations that the ultraviolet light is output; and
a communications interface, communicatively coupled to the controller, to transmit a report including at least one selected from the group consisting of: a path of operation of the autonomous mobile robot, and the one or more surfaces disinfected by the ultraviolet light.
2. The autonomous mobile robot of claim 1 , wherein the controller is configured to control an intensity of ultraviolet light output by the light source to disinfect the one or more surfaces within the area.
3. The autonomous mobile robot of claim 2 , wherein the controller is configured to track the intensity of the ultraviolet light that is output, and one or more locations that the ultraviolet light is output.
4. The autonomous mobile robot of claim 1 , wherein the controller is configured to turn on or off the light source to control the output of the ultraviolet light.
5. The autonomous mobile robot of claim 1 , wherein the disinfection of the one or more surfaces reduces the spread of at least one selected from the group consisting of: viruses, infectious diseases, bacteria, and organic microorganisms.
6. The autonomous mobile robot of claim 1 , wherein the at least one sensor comprises a three-dimensional camera that is disposed on at least one selected from the group consisting of: over the light source, and on the base.
7. The autonomous mobile robot of claim 1 , wherein the at least one sensor comprises a three-dimensional Light Detection and Ranging (LiDAR) sensor that is disposed on at least one selected from the group consisting of: over the light source, and on the base.
8. The autonomous mobile robot of claim 1 , wherein the at least one sensor comprises a first sensor and a second sensor, wherein the first sensor is disposed over the light source, and the second sensor is disposed on the base.
9. The autonomous mobile robot of claim 8 , wherein the light source comprises a plurality of bulbs disposed in an array, each having a first end and a second end,
wherein the first sensor is disposed over the first end of the bulbs, and the second end of the bulbs is disposed on the base and over the second sensor.
10. The autonomous mobile robot of claim 1 , wherein the at least one sensor is configured to detect a geometry of an object and determine that the object will impact the light source based on the detected geometry of the object and a path of movement of the mobile robot in the area and output a signal based on the determination.
11. The autonomous mobile robot of claim 10 , wherein the controller is configured to control the autonomous drive system to stop or change the path of the mobile robot before the light source impacts the object based on at least the signal received from the at least one sensor.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US18/111,200 US20230205211A1 (en) | 2020-03-16 | 2023-02-17 | Protection of ultraviolet (uv) light source on mobile device |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP20163352.6 | 2020-03-16 | ||
EP20163352.6A EP3882732B1 (en) | 2020-03-16 | 2020-03-16 | Protection of ultraviolet (uv) light source on mobile device |
US16/836,740 US11609570B2 (en) | 2020-03-16 | 2020-03-31 | Protection of ultraviolet (UV) light source on mobile device |
US18/111,200 US20230205211A1 (en) | 2020-03-16 | 2023-02-17 | Protection of ultraviolet (uv) light source on mobile device |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/836,740 Continuation US11609570B2 (en) | 2020-03-16 | 2020-03-31 | Protection of ultraviolet (UV) light source on mobile device |
Publications (1)
Publication Number | Publication Date |
---|---|
US20230205211A1 true US20230205211A1 (en) | 2023-06-29 |
Family
ID=69844694
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/836,740 Active 2041-01-29 US11609570B2 (en) | 2020-03-16 | 2020-03-31 | Protection of ultraviolet (UV) light source on mobile device |
US18/111,200 Pending US20230205211A1 (en) | 2020-03-16 | 2023-02-17 | Protection of ultraviolet (uv) light source on mobile device |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/836,740 Active 2041-01-29 US11609570B2 (en) | 2020-03-16 | 2020-03-31 | Protection of ultraviolet (UV) light source on mobile device |
Country Status (4)
Country | Link |
---|---|
US (2) | US11609570B2 (en) |
EP (1) | EP3882732B1 (en) |
DK (1) | DK3882732T3 (en) |
ES (1) | ES2939537T3 (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
USD1021055S1 (en) * | 2020-09-25 | 2024-04-02 | Xenex Disinfection Services Inc. | Oval cap assembly for a lighting apparatus |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6911177B2 (en) * | 2000-09-19 | 2005-06-28 | Jeffrey L. Deal | Ultraviolet area sterilizer and method of area sterilization using ultraviolet radiation |
US20040211444A1 (en) * | 2003-03-14 | 2004-10-28 | Taylor Charles E. | Robot vacuum with particulate detector |
US20080056933A1 (en) * | 2006-08-29 | 2008-03-06 | Moore Barrett H | Self-Propelled Sterilization Robot and Method |
US8779391B2 (en) * | 2011-03-03 | 2014-07-15 | Teckni-Corp | Sterilization system with ultraviolet emitter for eradicating biological contaminants |
US20120282135A1 (en) * | 2011-04-15 | 2012-11-08 | Samuel Richard Trapani | Room sterilization method and system |
US10010633B2 (en) * | 2011-04-15 | 2018-07-03 | Steriliz, Llc | Room sterilization method and system |
CN107865973B (en) | 2012-01-31 | 2020-08-04 | 瑟费塞德有限责任公司 | Hard surface disinfection system and method |
US10195298B2 (en) * | 2013-02-27 | 2019-02-05 | Arthur Kreitenberg | Internal sanitizing and communicating |
US20160271803A1 (en) * | 2013-05-03 | 2016-09-22 | Michael Stewart | Robotic disinfection system |
US10485887B2 (en) * | 2015-04-12 | 2019-11-26 | Angelica Holdings Llc | Targeted surface disinfection system with pulsed UV light |
WO2018132679A1 (en) * | 2017-01-12 | 2018-07-19 | UD Innovations, LLC | Portable uv-c disinfection apparatus, method, and system |
CN208641335U (en) | 2017-08-21 | 2019-03-26 | 西安富康空气净化设备工程有限公司 | A kind of movable type high energy xenon light sterilizing machine |
US10365364B1 (en) | 2018-05-18 | 2019-07-30 | Zendar Inc. | Systems and methods for detecting objects |
CN209529745U (en) | 2018-08-23 | 2019-10-25 | 上海钛米机器人科技有限公司 | Autonomous chlorination equipment |
-
2020
- 2020-03-16 EP EP20163352.6A patent/EP3882732B1/en active Active
- 2020-03-16 DK DK20163352.6T patent/DK3882732T3/en active
- 2020-03-16 ES ES20163352T patent/ES2939537T3/en active Active
- 2020-03-31 US US16/836,740 patent/US11609570B2/en active Active
-
2023
- 2023-02-17 US US18/111,200 patent/US20230205211A1/en active Pending
Also Published As
Publication number | Publication date |
---|---|
DK3882732T3 (en) | 2023-05-08 |
US11609570B2 (en) | 2023-03-21 |
US20210286363A1 (en) | 2021-09-16 |
EP3882732A1 (en) | 2021-09-22 |
ES2939537T3 (en) | 2023-04-24 |
EP3882732B1 (en) | 2023-02-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3888696B1 (en) | Method of plotting ultraviolet (uv) radiation for disinfection | |
US11602845B2 (en) | Method of detecting human and/or animal motion and performing mobile disinfection | |
US11701442B2 (en) | Method of autonomous mobile disinfection | |
JP5827891B2 (en) | Simultaneous positioning and map creation method | |
US20230205211A1 (en) | Protection of ultraviolet (uv) light source on mobile device | |
US20220288249A1 (en) | Method of self-disinfection of mobile robot | |
CN112462389A (en) | Mobile robot obstacle detection system, method and device and electronic equipment | |
KR102637851B1 (en) | Cleaning robot | |
US11687075B2 (en) | Remotely operated mobile service robots | |
TWI729733B (en) | Protection of ultraviolet (uv) light source on mobile device | |
KR20120113585A (en) | Intelligent linetracing robot and method for driving the same | |
CN113406950A (en) | Protection of ultraviolet light sources on mobile devices | |
US20230414813A1 (en) | Systems and methods of autonomous mobile disinfection with minimal input | |
US20230110302A1 (en) | Methods of sensor exposure validation | |
EP3952456B1 (en) | Methods of connecting to communications networks and switching network connectivity |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |