EP1537550A2 - Verfahren und vorrichtung zur implementierung eines mehrzweck überwachungssystems - Google Patents
Verfahren und vorrichtung zur implementierung eines mehrzweck überwachungssystemsInfo
- Publication number
- EP1537550A2 EP1537550A2 EP03764108A EP03764108A EP1537550A2 EP 1537550 A2 EP1537550 A2 EP 1537550A2 EP 03764108 A EP03764108 A EP 03764108A EP 03764108 A EP03764108 A EP 03764108A EP 1537550 A2 EP1537550 A2 EP 1537550A2
- Authority
- EP
- European Patent Office
- Prior art keywords
- photo
- objects
- pixel
- photographic
- processing
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19602—Image analysis to detect motion of the intruder, e.g. by frame subtraction
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19602—Image analysis to detect motion of the intruder, e.g. by frame subtraction
- G08B13/19604—Image analysis to detect motion of the intruder, e.g. by frame subtraction involving reference image or background adaptation with time to compensate for changing conditions, e.g. reference image update on detection of light level change
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19617—Surveillance camera constructional details
- G08B13/1963—Arrangements allowing camera rotation to change view, e.g. pivoting camera, pan-tilt and zoom [PTZ]
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19639—Details of the system layout
- G08B13/19641—Multiple cameras having overlapping views on a single scene
- G08B13/19643—Multiple cameras having overlapping views on a single scene wherein the cameras play different roles, e.g. different resolution, different camera type, master-slave camera
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19639—Details of the system layout
- G08B13/19647—Systems specially adapted for intrusion detection in or around a vehicle
- G08B13/1965—Systems specially adapted for intrusion detection in or around a vehicle the vehicle being an aircraft
-
- G—PHYSICS
- G08—SIGNALLING
- G08B—SIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
- G08B13/00—Burglar, theft or intruder alarms
- G08B13/18—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength
- G08B13/189—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems
- G08B13/194—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems
- G08B13/196—Actuation by interference with heat, light, or radiation of shorter wavelength; Actuation by intruding sources of heat, light, or radiation of shorter wavelength using passive radiation detection systems using image scanning and comparing systems using television cameras
- G08B13/19678—User interface
- G08B13/19691—Signalling events for better perception by user, e.g. indicating alarms by making display brighter, adding text, creating a sound
Definitions
- the present invention relates to the field of target detection system. More
- the invention relates to a method and apparatus for detecting a
- a foreign object in the area of airport runways may interfere with
- a foreign object can be a person, wildlife, birds, inanimate
- FOD such as birds, wildlife or any other object on the runway
- the means used for " deterring birds include vehicle/human . presence,
- JP 2,001,148,011 discloses a small animal detecting method and a small animal
- detecting device which can judge an intruder, a small animal, an insect, etc., by
- an image recognizing means on the basis of image data picked up by a camera.
- US 3,811,010 discloses an intrusion detection apparatus employing two spaced-
- comparator-adder analyzing circuitry is provided between the cameras and
- a radar system is used in order to detect and locate the location of
- dangerous objects may also not be natural ones, such as birds, but
- the method of the invention comprises the steps of: a) procuring, adjourning and storing in a memory files representing the
- the controlled space wherein said controlled space is
- dangerous parameters are the object size
- Said space may be divided into
- zones of different priorities viz. zones in which the observation is
- the method further comprises
- the method comprises documenting the data obtained from the observation of objects, for uture prevention acts.
- the future prevention acts are ehminating.
- the method of the present invention further comprises: a) generating
- said location being represented by the altitude, range and azimuth parameters of
- the imagers are cameras selected from the group consisting of: CCD
- CMOS based cameras or Forward Looking Infra Red (FLIR) cameras.
- FLIR Forward Looking Infra Red
- the apparatus according to the invention comprises:
- said devices can be one or more CCD
- CMOS camera and/or one or more Infra Red (IR) cameras;
- IR Infra Red
- the memory means may comprise a single or various electronic data storage
- the photographic devices are at least a pair of distinct and identical
- the apparatus According to a preferred embodiment of the present invention, the apparatus
- the elaborator means are one or more dedicated algorithms installed
- the apparatus is configured to control the computerized system. According to a preferred embodiment of the present invention, the apparatus
- a laser range finder which is electrically connected to the
- FIG. 1 schematically illustrates a monitoring system, according to a
- FIG. 2 schematically illustrates in a graph form a method of photographing
- Fig. 3 is a flow chart that shows the algorithm of a system for monitoring
- FIG. 4 schematically illustrates the data processing of the algorithm of Fig.
- FIG. 5 A schematically illustrates the detection of moving objects in the
- Fig. 5B schematically illustrates the detection of static objects in the data
- FIG. 6 schematically illustrates in a graph form the threshold level used for
- Fig. 7 schematically illustrates the solving of the general three
- Fig. 8 schematically illustrates a combined panoramic view and map
- FIG. 9 schematically illustrates a scanning of a sector around a vertical
- FIG. 10 schematically illustrates a scanning of a sector around a horizontal
- FIG. 11 schematically illustrates the monitoring system of Fig. 1 provided
- All the processing of this invention is digital processing. Taking a photograph by
- a camera or a digital camera such as those of the apparatus of this invention
- each pixel is associated a value that represents the radiation intensity value of
- the two- dimensional array of pixels therefore, is represented by a matrix consisting of an
- each digital or sampled image is provided with a corresponding digital or sampled image.
- the controlled space must be firstly
- the photographs must be taken generally include, e.g.,
- parameters may be, e.g., the size of the body, its apparent density, the
- the evaluation programs should be periodically updated, taking
- plan and elevation or each patrol at any time after an initial time.
- each patrol at any time after an initial time.
- the body is a living creature
- the documentation analysis can help to eliminate or reduce the
- Actions for eliminating the danger of collision Such actions may be carried out on the dangerous objects, and in that case they
- Fig. 1 schematically illustrates a monitoring system 10, according to a preferred
- System 10 comprises at least one photographic image
- CCD Charged Coupled Device
- Each photographic device can provide either color image or uncolored image.
- At least one of the photographic devices is a digital camera.
- each photographic device may have different type of
- each camera may be provided with lenses having different
- the photo raphic devices are used to allow
- the computerized system 15 is responsible for performing the processing
- computerized system 15 receives, at its inputs, data from active cameras that are
- system 10 e.g., CCD camera 11, thermal camera 12, CMOS based
- the data from the cameras is captured and digitized at the
- computerized system 15 processes the received data from the cameras in order to
- controller 151 controlled by controller 151 according to a set of instructions and data regarding
- system 15 outputs data regarding the detection of suspected dangerous, objects to
- monitors such as monitor 18, via its video card 17
- One or more of the cameras attached to system 10 is rotated by motors 13
- the reset sensor provides, to the computerized system
- the encoder provides, to the computerized system 15, the current angle of
- Motion controller 14 controls motors 13
- Motion controller 14 can be located
- Motion controller 14 communicates with the attached cameras and the
- the scanning is divided into several and a constant number of tracks, upon which each camera is
- the preferred scanning area is preformed at the area ground up to a
- the cameras of system 10 are installed on a
- the cameras can be configured in a variety of ways and positions. According to
- a pair of identical cameras is located
- distance between a pair of cameras is between 0.5 to 50 meter, horizontally,
- the cameras or imagers may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un-identical and may be un
- Fig. 2 schematically illustrates in a graph form an example for the method of
- the angle of the camera is modified before each photo or sequence of
- the camera takes the sequence of photos (shown by item 22) at a time period
- the additional details can be the distance of the object
- depth parameters are obtained (i.e., three-dimension
- At least two cameras it enables to elongate the detection range, as well as to
- between a pair of cameras is between 0.5 to 50 meter, the distance . can be
- Fig. 3 is a flow chart that shows an example of the program algorithm of system
- the initial definitions are parameters that are required for the operation of
- the initial camera angle definition regarding the area (such as, loading the lens
- next step, block 32 the computerized system 15 orders the motion controllers 14 to change the angle of the one or more camera.
- block 34 the next step
- the computerized system 15 orders the cameras (via motor controller 14) to take
- the sequence of photos preferably about 25 to 30 photos a second.
- the photos are
- step 33 The data processing in step 33 is performed in two stages.
- computerized system 15 decides whether a
- detected object is a dangerous object. If a dangerous object is detected, then at
- a warning signal is activated, such as showing the location of
- system 15 makes a decision that no dangerous body exists, then in the next step
- the last process data is stored in a related database.
- the stored data is used
- the background space is
- the data processing (block 33 of Fig. 3) is done in two stages.
- each pixel in each photo from the
- sequence of photos is mathematically processed from each camera that
- threshold value e.g., threshold 61 as shown
- the threshold value dynamically corresponds to the danger degrees.
- the pixels processing detects either moving objects or static objects, as
- objects are detected (i.e., pixels that their location on the Gaussian
- the 3-D data is used for detecting pixels that may
- a bird in a flock of birds may appear as a single pixel in the photo, but
- system 10 defines them as birds, even if
- each of the birds appears as a single pixel.
- system 10 find their location by
- dangerous object i.e., the suspected objects
- the measured parameters are compared to a predetermined table
- predetermined table of values is stored in memory 151 or other related
- the measured parameters can be: • 1. The dimension of the suspected object, its length and its width (e.g.,
- An object can be an adjacent group of pixels ' .
- Movement parameters such as direction that was created from one
- system 10 in case system 10 detects
- archive contains data and/or photos regarding the dangerous objects that were
- Fig. 5A schematically illustrates the detection of a moving object at the pixel
- Photo 42 is an average photo that was generated
- a comparison sequence of photos 451 to 480 is generated from the
- 451 to 480 represents the error value between photos 401 to 430 and
- Each error value is compared to a threshold level 61 (Fig. 6) in the
- the threshold level 61 is dynamically
- the location of the exceeded pixel is set to a specific
- a logic matrix 49 that represent the suspected photo e.g., the
- pixel is set as value of 255, wherein the other pixels value is set to 0).
- Fig. 5B schematically illustrates the detection of a static object at the pixel
- An average photo 42 is created from the current sequence of photos 401
- a derivative matrix 43 is generated from the average photo 42.
- derivative matrix 43 is used to emphasize relatively small objects in the
- the generated derivative matrix 43 is stored in a photo database 44
- the threshold level 61 is dynamically
- predetermined threshold level 61 the location of the exceeded pixel is
- the pixel is set as
- the generated logic matrix 49 that contains the suspected pixels is transferred to the logic process stage, wherein the suspicious pixels are
- the authorized bodies can be, for example, a navy
- system 10 is used for detecting burning in coal stratum.
- an IR camera such as those used by the present
- system 10 (Fig. 1)
- generating radiation i.e., a passive electro-optical radar.
- the location i.e., a passive electro-optical radar.
- polar coordinates e.g., range and azimuth.
- system 10 (Fig. 1) is used to measure and provide the
- a detected object such as the range, azimuth and altitude of the object.
- the location is relative to a reference coordinates system on earth. The location of the
- the imagers are digital photographic devices such as CCD or CMOS based
- FLIR Forward Looking Infra Red
- At least a pair of identical CCD cameras such as camera 12 of Fig. 1
- Each projection represents an
- each coordinate system has the pixel coordinate system (xl, yl) for the first camera and the pixel coordinate system (x2, y2) for the second camera (e.g., each coordinate system
- system 10 (Fig. 1) essentially comprises at least
- two cameras preferably having parallel optical axes and having synchronous
- a rotational motion means such as motor 13 (Fig. 1) and image
- the image processing means is used
- two cameras e.g., two units of CCD camera 12 (Fig. 1).
- Fig. 7 schematically illustrates the
- each scan step has a certain azimuth angle a which is
- ⁇ X X[ * cosa - Z t * since
- This embodiment further provides a passive operation of system 10 (Fig. 1) by imaging optical radiation in
- system 10 (Fig. 1) generates,' by elaborator means, a
- detected targets i.e., dangerous objects
- the elaborator means consisting of the
- Fig. 1 scans the monitored area by a vertical and/or horizontal
- the vertical rotational scanning is
- the horizontal rotational scanning is achieved by placing
- i ager means e.g., such as three or four CCD
- Fig. 8 schematically illustrates a combined panoramic view and map
- the electro-optical radar i.e., system
- the radar display is arranged in a graphical map presentation, 40, and a
- panoramic image 50 In the map, the relative locations of the targets, 60 and 70,
- the heights of the targets can be seen, while in the panoramic image, 50, the heights of the targets can be
- Open Graphic Library (OpenGL), as known to a skilled person in the art.
- additional video cameras e.g., CCD cameras
- CCD cameras operating in the normal vision band
- electro -optical radar i.e., system 10
- electro -optical radar i.e., system 10
- targets is measured by using radiation emitted or reflected from the target.
- location of the target is determined by using triangulation with the two cameras.
- This arrangement does not use active radiation emission from the radar itself
- the cameras are two system design parameters. As the distance between the two
- Each camera provides an image of the same area but from a different view or
- pixel in one image receives a vicinity of pixels in the other image
- detected targets may include all the measured features, e.g., target size,
- the present invention uses a panoramic image of the scene together with its map of detected targets to present
- FIG. 11 schematically illustrates the monitoring system of Fig. 1 provided with a
- laser range finder according to a preferred embodiment of the present invention.
- Laser Range Finder 200 is electrically connected to computerized system 15,
- the laser range is either via the CPU 152 and/or via the communication unit 19.
- finder 200 is used for measuring the distance of a detected object from it
- Laser Range Finder 200 preferably while system 10 monitors a given area.
- range finder 200 can be any suitable laser range finder device that may be fitted
- system 10 such as LDM 800-RS 232-WP industrial distance meter of
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Human Computer Interaction (AREA)
- Aviation & Aerospace Engineering (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
- Testing Or Calibration Of Command Recording Devices (AREA)
- Testing And Monitoring For Control Systems (AREA)
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IL15074502 | 2002-07-15 | ||
IL150745A IL150745A (en) | 2002-07-15 | 2002-07-15 | Method and apparatus for multipurpose monitoring system |
IL15381303 | 2003-01-06 | ||
IL15381303A IL153813A0 (en) | 2002-07-15 | 2003-01-06 | Method and apparatus for multipurpose monitoring system |
PCT/IL2003/000585 WO2004008403A2 (en) | 2002-07-15 | 2003-07-15 | Method and apparatus for implementing multipurpose monitoring system |
Publications (1)
Publication Number | Publication Date |
---|---|
EP1537550A2 true EP1537550A2 (de) | 2005-06-08 |
Family
ID=30117208
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP03764108A Ceased EP1537550A2 (de) | 2002-07-15 | 2003-07-15 | Verfahren und vorrichtung zur implementierung eines mehrzweck überwachungssystems |
Country Status (4)
Country | Link |
---|---|
US (1) | US8111289B2 (de) |
EP (1) | EP1537550A2 (de) |
AU (1) | AU2003242974A1 (de) |
WO (1) | WO2004008403A2 (de) |
Families Citing this family (97)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2414790A (en) * | 2004-06-04 | 2005-12-07 | Laser Optical Engineering Ltd | Detection of humans or animals by comparing infrared and visible light images |
US7864159B2 (en) | 2005-01-12 | 2011-01-04 | Thinkoptics, Inc. | Handheld vision based absolute pointing system |
IL168212A (en) | 2005-04-21 | 2012-02-29 | Rafael Advanced Defense Sys | System and method for protection of landed aircraft |
JP4773170B2 (ja) | 2005-09-14 | 2011-09-14 | 任天堂株式会社 | ゲームプログラムおよびゲームシステム |
US7851758B1 (en) * | 2005-09-29 | 2010-12-14 | Flir Systems, Inc. | Portable multi-function inspection systems and methods |
US20070121094A1 (en) * | 2005-11-30 | 2007-05-31 | Eastman Kodak Company | Detecting objects of interest in digital images |
US8913003B2 (en) * | 2006-07-17 | 2014-12-16 | Thinkoptics, Inc. | Free-space multi-dimensional absolute pointer using a projection marker system |
US9176598B2 (en) * | 2007-05-08 | 2015-11-03 | Thinkoptics, Inc. | Free-space multi-dimensional absolute pointer with improved performance |
BRPI0817039A2 (pt) | 2007-08-24 | 2015-07-21 | Stratech Systems Ltd | Sistema e método de vigilância de pista de pouso e decolagem |
DE102008018880A1 (de) * | 2008-04-14 | 2009-10-15 | Carl Zeiss Optronics Gmbh | Überwachungsverfahren und -vorrichtung für Windkraftanlagen, Gebäude mit transparenten Bereichen, Start- und Landebahnen und/oder Flugkorridore von Flughäfen |
EP2318804B1 (de) * | 2008-04-17 | 2017-03-29 | Shilat Optronics Ltd | Einbruchwarnsystem |
KR101733443B1 (ko) | 2008-05-20 | 2017-05-10 | 펠리칸 이매징 코포레이션 | 이종 이미저를 구비한 모놀리식 카메라 어레이를 이용한 이미지의 캡처링 및 처리 |
US11792538B2 (en) | 2008-05-20 | 2023-10-17 | Adeia Imaging Llc | Capturing and processing of images including occlusions focused on an image sensor by a lens stack array |
US8866920B2 (en) | 2008-05-20 | 2014-10-21 | Pelican Imaging Corporation | Capturing and processing of images using monolithic camera array with heterogeneous imagers |
FR2942062A1 (fr) * | 2009-02-12 | 2010-08-13 | Shaktiware | Systemes et procedes de detection d'objets |
DE102009016819B4 (de) | 2009-04-09 | 2011-12-15 | Carl Zeiss Optronics Gmbh | Verfahren zur Detektion wenigstens eines Objekts und/oder wenigstens einer Objektgruppe, Computerprogramm, Computerprogammprodukt, Stereokameraeinrichtung, aktiv Strahlung aussendendes Bildsensorsystem und Überwachungsvorrichtung |
FR2944934B1 (fr) * | 2009-04-27 | 2012-06-01 | Scutum | Procede et systeme de surveillance |
US8406925B2 (en) * | 2009-07-01 | 2013-03-26 | Honda Motor Co., Ltd. | Panoramic attention for humanoid robots |
TWI402777B (zh) * | 2009-08-04 | 2013-07-21 | Sinew System Tech Co Ltd | Management Method of Real Estate in Community Building |
US8514491B2 (en) | 2009-11-20 | 2013-08-20 | Pelican Imaging Corporation | Capturing and processing of images using monolithic camera array with heterogeneous imagers |
CN103004180A (zh) | 2010-05-12 | 2013-03-27 | 派力肯影像公司 | 成像器阵列和阵列照相机的架构 |
CN101916489A (zh) * | 2010-06-24 | 2010-12-15 | 北京华安天诚科技有限公司 | 机场跑道侵入告警服务器、***和方法 |
US8878950B2 (en) | 2010-12-14 | 2014-11-04 | Pelican Imaging Corporation | Systems and methods for synthesizing high resolution images using super-resolution processes |
WO2012115594A1 (en) * | 2011-02-21 | 2012-08-30 | Stratech Systems Limited | A surveillance system and a method for detecting a foreign object, debris, or damage in an airfield |
KR101973822B1 (ko) | 2011-05-11 | 2019-04-29 | 포토네이션 케이맨 리미티드 | 어레이 카메라 이미지 데이터를 송신 및 수신하기 위한 시스템들 및 방법들 |
US20120320151A1 (en) * | 2011-06-20 | 2012-12-20 | Howard Unger | Camera with automated panoramic image capture |
US8773501B2 (en) * | 2011-06-20 | 2014-07-08 | Duco Technologies, Inc. | Motorized camera with automated panoramic image capture sequences |
US20130265459A1 (en) | 2011-06-28 | 2013-10-10 | Pelican Imaging Corporation | Optical arrangements for use with an array camera |
US20130070060A1 (en) | 2011-09-19 | 2013-03-21 | Pelican Imaging Corporation | Systems and methods for determining depth from multiple views of a scene that include aliasing using hypothesized fusion |
JP6140709B2 (ja) | 2011-09-28 | 2017-05-31 | ペリカン イメージング コーポレイション | 明視野画像ファイルを符号化および復号するためのシステムおよび方法 |
US9412206B2 (en) | 2012-02-21 | 2016-08-09 | Pelican Imaging Corporation | Systems and methods for the manipulation of captured light field image data |
JP5753509B2 (ja) * | 2012-03-29 | 2015-07-22 | スタンレー電気株式会社 | 被測定物情報取得装置 |
US9210392B2 (en) | 2012-05-01 | 2015-12-08 | Pelican Imaging Coporation | Camera modules patterned with pi filter groups |
CN102707272B (zh) * | 2012-06-13 | 2014-03-19 | 西安电子科技大学 | 基于gpu的外辐射源雷达信号实时处理***及处理方法 |
KR20150023907A (ko) | 2012-06-28 | 2015-03-05 | 펠리칸 이매징 코포레이션 | 결함있는 카메라 어레이들, 광학 어레이들 및 센서들을 검출하기 위한 시스템들 및 방법들 |
US20140002674A1 (en) | 2012-06-30 | 2014-01-02 | Pelican Imaging Corporation | Systems and Methods for Manufacturing Camera Modules Using Active Alignment of Lens Stack Arrays and Sensors |
EP4296963A3 (de) | 2012-08-21 | 2024-03-27 | Adeia Imaging LLC | Verfahren zur tiefenerkennung in mit array-kameras aufgenommenen bildern |
CN104685513B (zh) | 2012-08-23 | 2018-04-27 | 派力肯影像公司 | 根据使用阵列源捕捉的低分辨率图像的基于特征的高分辨率运动估计 |
US9214013B2 (en) | 2012-09-14 | 2015-12-15 | Pelican Imaging Corporation | Systems and methods for correcting user identified artifacts in light field images |
US20140092281A1 (en) | 2012-09-28 | 2014-04-03 | Pelican Imaging Corporation | Generating Images from Light Fields Utilizing Virtual Viewpoints |
US9143711B2 (en) | 2012-11-13 | 2015-09-22 | Pelican Imaging Corporation | Systems and methods for array camera focal plane control |
US9091628B2 (en) | 2012-12-21 | 2015-07-28 | L-3 Communications Security And Detection Systems, Inc. | 3D mapping with two orthogonal imaging views |
US9462164B2 (en) | 2013-02-21 | 2016-10-04 | Pelican Imaging Corporation | Systems and methods for generating compressed light field representation data using captured light fields, array geometry, and parallax information |
US9253380B2 (en) | 2013-02-24 | 2016-02-02 | Pelican Imaging Corporation | Thin form factor computational array cameras and modular array cameras |
WO2014138697A1 (en) | 2013-03-08 | 2014-09-12 | Pelican Imaging Corporation | Systems and methods for high dynamic range imaging using array cameras |
US8866912B2 (en) | 2013-03-10 | 2014-10-21 | Pelican Imaging Corporation | System and methods for calibration of an array camera using a single captured image |
US9521416B1 (en) | 2013-03-11 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for image data compression |
US9888194B2 (en) | 2013-03-13 | 2018-02-06 | Fotonation Cayman Limited | Array camera architecture implementing quantum film image sensors |
WO2014164550A2 (en) | 2013-03-13 | 2014-10-09 | Pelican Imaging Corporation | System and methods for calibration of an array camera |
US9106784B2 (en) | 2013-03-13 | 2015-08-11 | Pelican Imaging Corporation | Systems and methods for controlling aliasing in images captured by an array camera for use in super-resolution processing |
US9519972B2 (en) | 2013-03-13 | 2016-12-13 | Kip Peli P1 Lp | Systems and methods for synthesizing images from image data captured by an array camera using restricted depth of field depth maps in which depth estimation precision varies |
US9100586B2 (en) | 2013-03-14 | 2015-08-04 | Pelican Imaging Corporation | Systems and methods for photometric normalization in array cameras |
US9578259B2 (en) | 2013-03-14 | 2017-02-21 | Fotonation Cayman Limited | Systems and methods for reducing motion blur in images or video in ultra low light with array cameras |
WO2014150856A1 (en) | 2013-03-15 | 2014-09-25 | Pelican Imaging Corporation | Array camera implementing quantum dot color filters |
EP2973476A4 (de) * | 2013-03-15 | 2017-01-18 | Pelican Imaging Corporation | Systeme und verfahren zur stereobildgebung mit kameraarrays |
US10122993B2 (en) | 2013-03-15 | 2018-11-06 | Fotonation Limited | Autofocus system for a conventional camera that uses depth information from an array camera |
US9633442B2 (en) | 2013-03-15 | 2017-04-25 | Fotonation Cayman Limited | Array cameras including an array camera module augmented with a separate camera |
US9497429B2 (en) | 2013-03-15 | 2016-11-15 | Pelican Imaging Corporation | Extended color processing on pelican array cameras |
US9445003B1 (en) | 2013-03-15 | 2016-09-13 | Pelican Imaging Corporation | Systems and methods for synthesizing high resolution images using image deconvolution based on motion and depth information |
US9898856B2 (en) | 2013-09-27 | 2018-02-20 | Fotonation Cayman Limited | Systems and methods for depth-assisted perspective distortion correction |
WO2015070105A1 (en) | 2013-11-07 | 2015-05-14 | Pelican Imaging Corporation | Methods of manufacturing array camera modules incorporating independently aligned lens stacks |
WO2015074078A1 (en) | 2013-11-18 | 2015-05-21 | Pelican Imaging Corporation | Estimating depth from projected texture using camera arrays |
US9456134B2 (en) | 2013-11-26 | 2016-09-27 | Pelican Imaging Corporation | Array camera configurations incorporating constituent array cameras and constituent cameras |
JP6329642B2 (ja) | 2013-12-10 | 2018-05-23 | エスゼット ディージェイアイ テクノロジー カンパニー リミテッドSz Dji Technology Co.,Ltd | センサ融合 |
US10089740B2 (en) | 2014-03-07 | 2018-10-02 | Fotonation Limited | System and methods for depth regularization and semiautomatic interactive matting using RGB-D images |
US9247117B2 (en) | 2014-04-07 | 2016-01-26 | Pelican Imaging Corporation | Systems and methods for correcting for warpage of a sensor array in an array camera module by introducing warpage into a focal plane of a lens stack array |
US9342884B2 (en) * | 2014-05-28 | 2016-05-17 | Cox Enterprises, Inc. | Systems and methods of monitoring waste |
US9521319B2 (en) | 2014-06-18 | 2016-12-13 | Pelican Imaging Corporation | Array cameras and array camera modules including spectral filters disposed outside of a constituent image sensor |
JP6181300B2 (ja) | 2014-09-05 | 2017-08-16 | エスゼット ディージェイアイ テクノロジー カンパニー リミテッドSz Dji Technology Co.,Ltd | 無人航空機の速度を制御するシステム |
WO2016033796A1 (en) | 2014-09-05 | 2016-03-10 | SZ DJI Technology Co., Ltd. | Context-based flight mode selection |
CN114675671A (zh) | 2014-09-05 | 2022-06-28 | 深圳市大疆创新科技有限公司 | 多传感器环境地图构建 |
US10250871B2 (en) | 2014-09-29 | 2019-04-02 | Fotonation Limited | Systems and methods for dynamic calibration of array cameras |
CN104536059B (zh) * | 2015-01-08 | 2017-03-08 | 西安费斯达自动化工程有限公司 | 图像/激光测距机场跑道异物监控一体化*** |
US9942474B2 (en) | 2015-04-17 | 2018-04-10 | Fotonation Cayman Limited | Systems and methods for performing high speed video capture and depth estimation using array cameras |
US9906733B2 (en) * | 2015-06-22 | 2018-02-27 | The Johns Hopkins University | Hardware and system for single-camera stereo range determination |
JP6450852B2 (ja) * | 2015-09-17 | 2019-01-09 | 株式会社日立国際電気 | 落下物検知追跡システム |
WO2017153979A1 (en) | 2016-03-06 | 2017-09-14 | Foresight Automotive Ltd. | Running vehicle alerting system and method |
EP3261071B1 (de) * | 2016-06-22 | 2020-04-01 | Outsight | Verfahren und system für das detektieren von eindringungen eines überwachten volumens |
CN106597556B (zh) * | 2016-12-09 | 2019-01-15 | 北京无线电计量测试研究所 | 一种机场跑道异物检测***背景对消的方法 |
US10482618B2 (en) | 2017-08-21 | 2019-11-19 | Fotonation Limited | Systems and methods for hybrid depth regularization |
US11436823B1 (en) | 2019-01-21 | 2022-09-06 | Cyan Systems | High resolution fast framing infrared detection system |
CN109751962A (zh) * | 2019-03-11 | 2019-05-14 | 冀中能源峰峰集团有限公司 | 一种基于机器视觉的煤体积动态计量装置及方法 |
US11448483B1 (en) | 2019-04-29 | 2022-09-20 | Cyan Systems | Projectile tracking and 3D traceback method |
WO2021061245A2 (en) | 2019-06-28 | 2021-04-01 | Cyan Systems | Fast framing moving target imaging system and method |
WO2021055585A1 (en) | 2019-09-17 | 2021-03-25 | Boston Polarimetrics, Inc. | Systems and methods for surface modeling using polarization cues |
US11525906B2 (en) | 2019-10-07 | 2022-12-13 | Intrinsic Innovation Llc | Systems and methods for augmentation of sensor systems and imaging systems with polarization |
KR20230116068A (ko) | 2019-11-30 | 2023-08-03 | 보스턴 폴라리메트릭스, 인크. | 편광 신호를 이용한 투명 물체 분할을 위한 시스템및 방법 |
WO2021154386A1 (en) | 2020-01-29 | 2021-08-05 | Boston Polarimetrics, Inc. | Systems and methods for characterizing object pose detection and measurement systems |
US11797863B2 (en) | 2020-01-30 | 2023-10-24 | Intrinsic Innovation Llc | Systems and methods for synthesizing data for training statistical models on different imaging modalities including polarized images |
US11953700B2 (en) | 2020-05-27 | 2024-04-09 | Intrinsic Innovation Llc | Multi-aperture polarization optical systems using beam splitters |
CN112668461B (zh) * | 2020-12-25 | 2023-05-23 | 浙江弄潮儿智慧科技有限公司 | 一种具有野生动物识别的智能监管*** |
US12020455B2 (en) | 2021-03-10 | 2024-06-25 | Intrinsic Innovation Llc | Systems and methods for high dynamic range image reconstruction |
US11290658B1 (en) | 2021-04-15 | 2022-03-29 | Boston Polarimetrics, Inc. | Systems and methods for camera exposure control |
US11954886B2 (en) | 2021-04-15 | 2024-04-09 | Intrinsic Innovation Llc | Systems and methods for six-degree of freedom pose estimation of deformable objects |
CN113033481B (zh) * | 2021-04-20 | 2023-06-02 | 湖北工业大学 | 基于一阶全卷积目标检测算法的手持棍棒的检测方法 |
US11689813B2 (en) | 2021-07-01 | 2023-06-27 | Intrinsic Innovation Llc | Systems and methods for high dynamic range imaging using crossed polarizers |
CN114063641B (zh) * | 2021-10-19 | 2024-04-16 | 深圳市优必选科技股份有限公司 | 机器人巡逻方法、巡逻机器人及计算机可读存储介质 |
Family Cites Families (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3811010A (en) | 1972-08-16 | 1974-05-14 | Us Navy | Intrusion detection apparatus |
US4429328A (en) | 1981-07-16 | 1984-01-31 | Cjm Associates | Three-dimensional display methods using vertically aligned points of origin |
FR2641871B1 (fr) | 1989-01-18 | 1991-07-26 | Telecommunications Sa | Systeme de determination de la position d'au moins une cible par triangulation |
US5175616A (en) * | 1989-08-04 | 1992-12-29 | Her Majesty The Queen In Right Of Canada, As Represented By The Minister Of National Defence Of Canada | Stereoscopic video-graphic coordinate specification system |
US4989084A (en) | 1989-11-24 | 1991-01-29 | Wetzel Donald C | Airport runway monitoring system |
DE4113992A1 (de) * | 1991-04-29 | 1992-11-05 | Ameling Walter | Verfahren zur automatischen dreidimensionalen ueberwachung von gefahrenraeumen |
ES2049176B1 (es) | 1992-08-07 | 1997-07-01 | J P Producciones S L 50 | Sistema de filmacion estereoscopica-monoscopica con grabacion de hasta 360 grados en vertical, y camara de objetivos rotatorios correspondientes. |
US5666157A (en) | 1995-01-03 | 1997-09-09 | Arc Incorporated | Abnormality detection and surveillance system |
JP3569992B2 (ja) * | 1995-02-17 | 2004-09-29 | 株式会社日立製作所 | 移動体検出・抽出装置、移動体検出・抽出方法及び移動体監視システム |
US5790183A (en) * | 1996-04-05 | 1998-08-04 | Kerbyson; Gerald M. | High-resolution panoramic television surveillance system with synoptic wide-angle field of view |
US5686889A (en) * | 1996-05-20 | 1997-11-11 | The United States Of America As Represented By The Secretary Of The Army | Infrared sniper detection enhancement |
US5953054A (en) * | 1996-05-31 | 1999-09-14 | Geo-3D Inc. | Method and system for producing stereoscopic 3-dimensional images |
DE19621612C2 (de) * | 1996-05-31 | 2001-03-01 | C Vis Comp Vision Und Automati | Vorrichtung zur Überwachung eines Gleisabschnittes in einem Bahnhof |
US6724931B1 (en) * | 1996-12-02 | 2004-04-20 | Hsu Shin-Yi | Compilable plain english-like language for extracting objects from an image using a primitive image map |
US6113343A (en) * | 1996-12-16 | 2000-09-05 | Goldenberg; Andrew | Explosives disposal robot |
DE19709799A1 (de) | 1997-03-10 | 1998-09-17 | Bosch Gmbh Robert | Einrichtung zur Videoüberwachung einer Fläche |
EP0878965A3 (de) | 1997-05-14 | 2000-01-12 | Hitachi Denshi Kabushiki Kaisha | Verfahrn zur Verfolgung eines eintretenden Objektes und Gerät zu dessen Verfolgung und Überwachung |
DE19809210A1 (de) * | 1998-03-04 | 1999-09-16 | Siemens Ag | Verfahren und Vorrichtung zur Überwachung einer Szene |
JP3779494B2 (ja) * | 1998-06-03 | 2006-05-31 | 松下電器産業株式会社 | 動き検出装置及び記録媒体 |
US6512537B1 (en) | 1998-06-03 | 2003-01-28 | Matsushita Electric Industrial Co., Ltd. | Motion detecting apparatus, motion detecting method, and storage medium storing motion detecting program for avoiding incorrect detection |
US6970183B1 (en) * | 2000-06-14 | 2005-11-29 | E-Watch, Inc. | Multimedia surveillance and monitoring system including network configuration |
US6023588A (en) | 1998-09-28 | 2000-02-08 | Eastman Kodak Company | Method and apparatus for capturing panoramic images with range data |
JP2001148011A (ja) | 1999-11-19 | 2001-05-29 | Fujitsu General Ltd | 画像認識による小動物識別方法および小動物識別装置 |
DE10032433A1 (de) * | 2000-07-04 | 2002-01-17 | H A N D Gmbh | Verfahren zur Bodenraumüberwachung |
DE10049366A1 (de) | 2000-10-05 | 2002-04-25 | Ind Technik Ips Gmbh | Verfahren zum Überwachen eines Sicherheitsbereichs und entsprechendes System |
US6954498B1 (en) * | 2000-10-24 | 2005-10-11 | Objectvideo, Inc. | Interactive video manipulation |
-
2003
- 2003-07-15 WO PCT/IL2003/000585 patent/WO2004008403A2/en not_active Application Discontinuation
- 2003-07-15 US US10/521,207 patent/US8111289B2/en not_active Expired - Fee Related
- 2003-07-15 EP EP03764108A patent/EP1537550A2/de not_active Ceased
- 2003-07-15 AU AU2003242974A patent/AU2003242974A1/en not_active Abandoned
Non-Patent Citations (1)
Title |
---|
See references of WO2004008403A3 * |
Also Published As
Publication number | Publication date |
---|---|
AU2003242974A8 (en) | 2004-02-02 |
WO2004008403A3 (en) | 2004-03-11 |
US8111289B2 (en) | 2012-02-07 |
AU2003242974A1 (en) | 2004-02-02 |
WO2004008403A2 (en) | 2004-01-22 |
US20060049930A1 (en) | 2006-03-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8111289B2 (en) | Method and apparatus for implementing multipurpose monitoring system | |
Hammer et al. | Lidar-based detection and tracking of small UAVs | |
US9420177B2 (en) | Panoramic view imaging system with laser range finding and blind spot detection | |
JP6110316B2 (ja) | 監視システムおよび飛行場内の異物、破片、または損害を検出する方法 | |
CN111679695B (zh) | 一种基于深度学习技术的无人机巡航及追踪***和方法 | |
WO2011060385A1 (en) | Method for tracking an object through an environment across multiple cameras | |
Bhadwal et al. | Smart border surveillance system using wireless sensor network and computer vision | |
CN112068111A (zh) | 一种基于多传感器信息融合的无人机目标侦测方法 | |
US11335026B1 (en) | Detecting target objects in a 3D space | |
Hammer et al. | Potential of lidar sensors for the detection of UAVs | |
Hammer et al. | UAV detection, tracking, and classification by sensor fusion of a 360 lidar system and an alignable classification sensor | |
CN111899447A (zh) | 监控***和方法 | |
US11823550B2 (en) | Monitoring device and method for monitoring a man-overboard in a ship section | |
CN208768193U (zh) | 近空智能监控*** | |
CN112802100A (zh) | 一种入侵检测方法、装置、设备和计算机可读存储介质 | |
Müller et al. | Drone detection, recognition, and assistance system for counter-UAV with VIS, radar, and radio sensors | |
CN108769628A (zh) | 近空智能监控***及方法 | |
Lohani et al. | Surveillance system based on Flash LiDAR | |
IL153813A (en) | Method and apparatus for multipurpose monitoring system | |
US20180010911A1 (en) | Ground-Based System for Geolocation of Perpetrators of Aircraft Laser Strikes | |
RU147594U1 (ru) | Устройство голографической скрытности объектов от малогабаритных беспилотных летательных аппаратов | |
Titov et al. | Multispectral optoelectronic device for controlling an autonomous mobile platform | |
US20230342952A1 (en) | Method for coordinative measuring by terrestrial scanning with image-based interference detection of moving objects | |
Tulldahl et al. | Application and capabilities of lidar from small UAV | |
Arlowe et al. | The mobile intrusion detection and assessment system (MIDAS) |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20050211 |
|
AK | Designated contracting states |
Kind code of ref document: A2 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LI LU MC NL PT RO SE SI SK TR |
|
17Q | First examination report despatched |
Effective date: 20051019 |
|
APBN | Date of receipt of notice of appeal recorded |
Free format text: ORIGINAL CODE: EPIDOSNNOA2E |
|
APBR | Date of receipt of statement of grounds of appeal recorded |
Free format text: ORIGINAL CODE: EPIDOSNNOA3E |
|
APAF | Appeal reference modified |
Free format text: ORIGINAL CODE: EPIDOSCREFNE |
|
APBT | Appeal procedure closed |
Free format text: ORIGINAL CODE: EPIDOSNNOA9E |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION HAS BEEN REFUSED |
|
18R | Application refused |
Effective date: 20080115 |