US20130141526A1 - Apparatus and Method for Video Image Stitching - Google Patents
Apparatus and Method for Video Image Stitching Download PDFInfo
- Publication number
- US20130141526A1 US20130141526A1 US13/691,632 US201213691632A US2013141526A1 US 20130141526 A1 US20130141526 A1 US 20130141526A1 US 201213691632 A US201213691632 A US 201213691632A US 2013141526 A1 US2013141526 A1 US 2013141526A1
- Authority
- US
- United States
- Prior art keywords
- video camera
- video
- image
- parameters
- camera
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H04N5/23238—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/4728—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for selecting a Region Of Interest [ROI], e.g. for requesting a higher resolution version of a selected region
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/698—Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/21—Server components or server architectures
- H04N21/218—Source of audio or video content, e.g. local disk arrays
- H04N21/21805—Source of audio or video content, e.g. local disk arrays enabling multiple viewpoints, e.g. using a plurality of cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
- H04N21/234363—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the spatial resolution, e.g. for clients with a lower screen resolution
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
- H04N21/23439—Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements for generating different versions
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/239—Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests
- H04N21/2393—Interfacing the upstream path of the transmission network, e.g. prioritizing client content requests involving handling client requests
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/422—Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
- H04N21/4223—Cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/45—Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
- H04N21/466—Learning process for intelligent management, e.g. learning user preferences for recommending movies
- H04N21/4668—Learning process for intelligent management, e.g. learning user preferences for recommending movies for recommending content, e.g. movies
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/47202—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for requesting content on demand, e.g. video on demand
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/816—Monomedia components thereof involving special video data, e.g 3D video
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2101/00—Still video cameras
Definitions
- This invention relates generally to video image processing. More particularly, this invention relates to panoramic video image stitching utilizing alignment and calibration parameters for dynamic video image compensation.
- Panoramic video feeds should be combined in a seamless manner.
- Existing techniques that endeavor to achieve this have high power requirements and excessive processing times, particularly in the context of mobile devices, which are constrained by these parameters.
- a method includes collecting a first image from a first video camera of a test pattern and a second image from a second video camera of the test pattern.
- the first image and the second image have an overlap region.
- the overlap region is evaluated to generate calibration parameters that accommodate for any vertical, horizontal or rotational misalignment between the first image and the second image.
- Calibration parameters are applied to video streams from the first video camera and the second video camera.
- FIG. 1 illustrates a calibration system utilized in accordance with an embodiment of the invention.
- FIG. 2 illustrates processing operations associated with an embodiment of the invention.
- FIGS. 3 and 4 illustrate image correction operations performed in accordance with an embodiment of the invention.
- FIGS. 5 and 6 illustrate error correction operations performed in accordance with an embodiment of the invention.
- An embodiment of the invention enables image stitching at high frame rates to produce an output video data stream created from more than one input video stream.
- Test and calibration processes using still photographs of charts and video capture of moving targets enable alignment of the cameras and determine the known overlap between cameras.
- FIG. 1 illustrates a calibration technique utilized in accordance with an embodiment of the invention.
- a first camera 100 with a viewing angle 102 takes an image of a test pattern 104 .
- a second camera 106 with a viewing angle 108 takes an image of the test pattern 104 .
- the viewing angles have an overlap region 110 .
- the cameras 100 and 106 have fixed positions in mount 112 .
- Each camera delivers image data to an image processor 114 .
- the image processor 114 may be a computer with a central processor, graphics processing unit and/or an additional processor.
- Still photographs are used to determine the fixed overlap area 110 from one image sensor to the adjacent image sensor based on alignment within the final system, or based on a standardized calibration fixture.
- Sharp lines of the test pattern 104 allow the image processor 144 to program the known overlap at the pixel level. That is, the image processor 114 accommodates for any vertical, horizontal or rotational misalignment.
- the test pattern 104 also allows the image processor 114 to determine any focus errors or areas of soft focus for each image sensor so that image correction processing can be applied later to produce uniformly sharp images.
- the test pattern 104 may be in the form of a grid.
- the field of view of each image capture device will have areas where the image has some distortion caused by either the lens or some other irregularity in the system.
- the images should appear to have no distortion or uniform distortion along the edge where the input streams are joined.
- the test pattern 104 may also be in the form of a color chart and/or uniform grey chart. Such a test pattern allows the image processor 114 to analyze any potential differences in color accuracy, relative illumination, and relative uniformity between cameras. These differences are stored as correction factors utilized by the image processor 114 to reduce noticeable differences between image streams.
- This calibration process allows for an output video stream to be assembled from multiple video streams with the viewer being unable to perceive any meaningful change in image quality throughout an entire field of view, including up to 360 degrees along multiple axes.
- FIG. 2 illustrates processing operations associated with an embodiment of the invention.
- the image processor 114 simultaneously captures frames from the first camera 100 and the second camera 106 , as shown with block 200 .
- Stored known camera distortion parameters 202 are applied by the image processor 204 .
- the frames may then be converted from rectangular coordinates to cylindrical or spherical coordinates 206 . It is then determined whether alignment has been calculated 208 . If not, alignment is calculated 210 and the alignment calculation is stored 212 and processing proceeds to block 214 .
- adjacent frames are stitched and blended based on the alignment calculation 214 . Adjacent frames are then applied as output 216 to a display.
- the following code implements operations of FIG. 2 .
- the process of automated alignment uses processing techniques implemented in OpenCV to determine the offset between two images that have overlapping regions. For example, feature matching and homography calculations may be used.
- manual alignment may be used to manually position image frames in a panoramic frame until the overlapping regions are seamless.
- the beginning and end of the frame need to be aligned and stitched together. To accomplish this there are three parameters that need to be calibrated.
- the invention may be used in the production of video camera systems that create ultra wide-angle video streams or panoramic video.
- the invention can also be used post-production to address camera performance issues in the field. If cameras are knocked out of alignment during use, the calibration and alignment steps can re-calibrate the system, preventing the unnecessary replacement of certain parts or the whole system.
- This invention can also be used when creating camera systems in multiple axes. For example, if camera systems are stacked along a vertical axis to create a “tall” cylindrical video stream, the alignment and calibration process can be used to ensure that camera performance is set within a pre-defined standard so there is not a wide variation in image performance between the separate video camera systems.
- the invention also allows multiple cameras to be calibrated within a specific performance range. If there are multiple camera systems being used at the same event, it is critical that they all have image performance within a defined range to reduce variation in image performance/quality between camera systems.
- the invention allows for calibration across a group of panoramic or wide-angle camera systems to ensure video outputs appear consistent to the end user.
- a reduction in bandwidth can be achieved by calibrating the Lens Distortion Parameter and the Camera Field of View (FOV) for each camera in the system.
- FOV Camera Field of View
- the system processes each camera's video stream frame by frame it will use the Lens Distortion Parameter and Camera FOV to correct the inherent lens distortion/perspective in each frame. Correcting the video frames with these parameters warps the pixels to a certain degree, causing the frame to be cropped in order to get rid of the warping affects.
- FIG. 3 illustrates distortion 300 between individual images 302 and 304 .
- cropping lines 400 and 402 may be used to compensate for this warping affect and provide a more uniform viewing area that is pleasing to the eye.
- the image processor 114 may execute software with calibrated parameters that reduce the overall time to stitch camera frames together. These parameters may include FOV overlap, image alignment, individual camera illumination correction and system relative illumination.
- Calibrating the image alignment and exposure values for the system can reduce the overall per-frame system processing time by approximately ⁇ 400 ms and ⁇ 450 ms respectively. This improvement is critical for live viewing applications, or applications when the camera may be turned on and off frequently.
- Image alignment is the parameter that defines the offset between each camera in order to stitch and produce a complete 360° panoramic video.
- Each camera's exposure settings should be set to be the same and change at the same rate to eliminate under-exposed and over-exposed areas in the 360-degree image.
- Dynamic range software can help compensate for cameras that are under or over exposed during difficult lighting conditions when the system is set to a single exposure level.
- the alignment and calibration techniques may be used for initial system set up and initial video streaming. Further, they may be used once the system is operative and is otherwise streaming video. In such a context, real world problems occur, such as camera faults and subsequent camera displacement or offset. Techniques of the invention address these issues.
- FIG. 5 illustrates processing associated with one such technique.
- the image processor 114 streams video from multiple cameras 500 .
- the image processor 114 is configured to check for a faulty camera 502 (e.g., a lost signal). If a faulty camera exists, the image processor 114 maximizes the field of view of adjacent cameras 504 . As shown in FIG. 1 , an overlapping region 110 exists between cameras. So for example, if camera 100 is faulty, the full field of view 108 of camera 106 would be utilized.
- FIG. 6 illustrates processing associated with a camera misalignment after initial alignment and calibration.
- the image processor 114 streams video from multiple cameras 600 .
- the image processor 114 is configured to check for an offset camera 602 (e.g., a camera misaligned from its original position). If an offset camera is identified, the offset is evaluated 604 . For example, the displacement between matching pixels of adjacent video streams may be evaluated. Also, outside sensors may be used to determine this displacement, such as accelerometers or digital gyroscopes. This displacement is then used to form updated alignment parameters (e.g., the original alignment parameters are compensated to incorporate the identified displacement) 606 . Video streams are subsequently stitched based on the updated alignment parameters 608 .
- an offset camera 602 e.g., a camera misaligned from its original position.
- the offset is evaluated 604 . For example, the displacement between matching pixels of adjacent video streams may be evaluated. Also, outside sensors may be used to determine this displacement, such as accelerometers or digital gy
- An embodiment of the present invention relates to a computer storage product with a computer readable storage medium having computer code thereon for performing various computer-implemented operations.
- the media and computer code may be those specially designed and constructed for the purposes of the present invention, or they may be of the kind well known and available to those having skill in the computer software arts.
- Examples of computer-readable media include, but are not limited to: magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROMs, DVDs and holographic devices; magneto-optical media; and hardware devices that are specially configured to store and execute program code, such as application-specific integrated circuits (“ASICs”), programmable logic devices (“PLDs”) and ROM and RAM devices.
- ASICs application-specific integrated circuits
- PLDs programmable logic devices
- Examples of computer code include machine code, such as produced by a compiler, and files containing higher-level code that are executed by a computer using an interpreter.
- machine code such as produced by a compiler
- files containing higher-level code that are executed by a computer using an interpreter.
- an embodiment of the invention may be implemented using JAVA®, C++, or other object-oriented programming language and development tools.
- Another embodiment of the invention may be implemented in hardwired circuitry in place of, or in combination with, machine-executable software instructions.
Abstract
Description
- This application claims priority to U.S. Provisional Patent Application 61/566,269, filed Dec. 2, 2011, entitled “Panoramic video Camera System and Related Methods”, the contents of which are incorporated herein by reference.
- This invention relates generally to video image processing. More particularly, this invention relates to panoramic video image stitching utilizing alignment and calibration parameters for dynamic video image compensation.
- Panoramic video feeds should be combined in a seamless manner. Existing techniques that endeavor to achieve this have high power requirements and excessive processing times, particularly in the context of mobile devices, which are constrained by these parameters.
- A method includes collecting a first image from a first video camera of a test pattern and a second image from a second video camera of the test pattern. The first image and the second image have an overlap region. The overlap region is evaluated to generate calibration parameters that accommodate for any vertical, horizontal or rotational misalignment between the first image and the second image. Calibration parameters are applied to video streams from the first video camera and the second video camera.
- The invention is more fully appreciated in connection with the following detailed description taken in conjunction with the accompanying drawings, in which:
-
FIG. 1 illustrates a calibration system utilized in accordance with an embodiment of the invention. -
FIG. 2 illustrates processing operations associated with an embodiment of the invention. -
FIGS. 3 and 4 illustrate image correction operations performed in accordance with an embodiment of the invention. -
FIGS. 5 and 6 illustrate error correction operations performed in accordance with an embodiment of the invention. - Like reference numerals refer to corresponding parts throughout the several views of the drawings.
- An embodiment of the invention enables image stitching at high frame rates to produce an output video data stream created from more than one input video stream. Test and calibration processes using still photographs of charts and video capture of moving targets enable alignment of the cameras and determine the known overlap between cameras.
-
FIG. 1 illustrates a calibration technique utilized in accordance with an embodiment of the invention. Afirst camera 100 with aviewing angle 102 takes an image of atest pattern 104. Similarly, asecond camera 106 with aviewing angle 108 takes an image of thetest pattern 104. The viewing angles have anoverlap region 110. Thecameras mount 112. Each camera delivers image data to animage processor 114. Theimage processor 114 may be a computer with a central processor, graphics processing unit and/or an additional processor. - Still photographs are used to determine the fixed
overlap area 110 from one image sensor to the adjacent image sensor based on alignment within the final system, or based on a standardized calibration fixture. Sharp lines of thetest pattern 104 allow the image processor 144 to program the known overlap at the pixel level. That is, theimage processor 114 accommodates for any vertical, horizontal or rotational misalignment. Thetest pattern 104 also allows theimage processor 114 to determine any focus errors or areas of soft focus for each image sensor so that image correction processing can be applied later to produce uniformly sharp images. - The
test pattern 104 may be in the form of a grid. In this case, the field of view of each image capture device will have areas where the image has some distortion caused by either the lens or some other irregularity in the system. To produce an output video stream from more than one video input, the images should appear to have no distortion or uniform distortion along the edge where the input streams are joined. By calculating a known distortion for each image capture device, distortion for individual cameras is corrected. Since resulting distortions for points in the image plane will be known for each image capture device, the distortion between images will be corrected by image processing to create an image with minimal distortion between camera data streams. - The
test pattern 104 may also be in the form of a color chart and/or uniform grey chart. Such a test pattern allows theimage processor 114 to analyze any potential differences in color accuracy, relative illumination, and relative uniformity between cameras. These differences are stored as correction factors utilized by theimage processor 114 to reduce noticeable differences between image streams. - This calibration process allows for an output video stream to be assembled from multiple video streams with the viewer being unable to perceive any meaningful change in image quality throughout an entire field of view, including up to 360 degrees along multiple axes.
-
FIG. 2 illustrates processing operations associated with an embodiment of the invention. Theimage processor 114 simultaneously captures frames from thefirst camera 100 and thesecond camera 106, as shown withblock 200. Stored knowncamera distortion parameters 202 are applied by theimage processor 204. The frames may then be converted from rectangular coordinates to cylindrical orspherical coordinates 206. It is then determined whether alignment has been calculated 208. If not, alignment is calculated 210 and the alignment calculation is stored 212 and processing proceeds to block 214. Atblock 214, adjacent frames are stitched and blended based on thealignment calculation 214. Adjacent frames are then applied asoutput 216 to a display. - Thus, through the calibration process, camera distortion and alignment parameters are obtained. These parameters are subsequently utilized as additional frames are received.
- The following is an example of lens distortion correction and perspective correction applied to an image in accordance with an embodiment of the invention. Consider the following parameters.
-
- Image Size: 1280 px×720 px
- Image Center: (640, 360)
- Distortion Size: 1 (Scaling factor, 0-1)
- Distortion: 0.03 (Distortion correction polynomial coefficient)
- Distant Param: 103.00 px (Focal length in pixels)
- The following code implements operations of
FIG. 2 . -
vec2 tc = gl_TexCoord[0].st; // Lens Distortion Correction vec2 P = (tc / imageCenter) − 1.; // to normalized image coordinates P /= distortionSize; vec2 Pp = P / (1. − distortion * dot(P, P)); P *= distortionSize; tc = (Pp + 1.) * imageCenter; // back to pixel coordinates // Cartesian coordinates tc −= vec2( imageCenter.x, imageCenter.y ); // Sphere(FishEye)-to-Erectangular float phi = tc.s / distanceparam; float theta = −tc.t / distanceparam + (M_PI/2.); if (theta < 0.0) { theta = −theta; phi += M_PI; } if (theta > M_PI) { theta = M_PI − (theta − M_PI); phi += M_PI ; } float s = sin(theta); vec2 v = vec2(s * sin(phi), cos(theta)); float r = length(v); theta = distanceparam * atan(r, s * cos(phi)); tc = v * (theta / r); // Erectangular-to-Cylindrical tc.t = distanceparam * atan(tc.t / distanceparam ); //Pixel coordinates tc += vec2( imageCenter.x − 0.5, imageCenter.y − 0.5); - In one embodiment, the process of automated alignment uses processing techniques implemented in OpenCV to determine the offset between two images that have overlapping regions. For example, feature matching and homography calculations may be used.
- In one embodiment, manual alignment may be used to manually position image frames in a panoramic frame until the overlapping regions are seamless.
When dealing with a 360-degree panoramic frame, the beginning and end of the frame need to be aligned and stitched together. To accomplish this there are three parameters that need to be calibrated. -
- Overlap—The amount of pixel overlap between the beginning and end of the frame.
- Twist—The amount of vertical error between the beginning and end of the frame.
- Blend Width—The width of the blending area across the overlapping region.
- Those skilled in the art will appreciate that the invention may be used in the production of video camera systems that create ultra wide-angle video streams or panoramic video. The invention can also be used post-production to address camera performance issues in the field. If cameras are knocked out of alignment during use, the calibration and alignment steps can re-calibrate the system, preventing the unnecessary replacement of certain parts or the whole system.
- This invention can also be used when creating camera systems in multiple axes. For example, if camera systems are stacked along a vertical axis to create a “tall” cylindrical video stream, the alignment and calibration process can be used to ensure that camera performance is set within a pre-defined standard so there is not a wide variation in image performance between the separate video camera systems.
- The invention also allows multiple cameras to be calibrated within a specific performance range. If there are multiple camera systems being used at the same event, it is critical that they all have image performance within a defined range to reduce variation in image performance/quality between camera systems. The invention allows for calibration across a group of panoramic or wide-angle camera systems to ensure video outputs appear consistent to the end user.
- A reduction in bandwidth can be achieved by calibrating the Lens Distortion Parameter and the Camera Field of View (FOV) for each camera in the system. When the system processes each camera's video stream frame by frame it will use the Lens Distortion Parameter and Camera FOV to correct the inherent lens distortion/perspective in each frame. Correcting the video frames with these parameters warps the pixels to a certain degree, causing the frame to be cropped in order to get rid of the warping affects. For example,
FIG. 3 illustratesdistortion 300 betweenindividual images FIG. 4 , croppinglines - The
image processor 114 may execute software with calibrated parameters that reduce the overall time to stitch camera frames together. These parameters may include FOV overlap, image alignment, individual camera illumination correction and system relative illumination. - Calibrating the image alignment and exposure values for the system can reduce the overall per-frame system processing time by approximately ˜400 ms and ˜450 ms respectively. This improvement is critical for live viewing applications, or applications when the camera may be turned on and off frequently.
- Image alignment is the parameter that defines the offset between each camera in order to stitch and produce a complete 360° panoramic video. Each camera's exposure settings should be set to be the same and change at the same rate to eliminate under-exposed and over-exposed areas in the 360-degree image. Dynamic range software can help compensate for cameras that are under or over exposed during difficult lighting conditions when the system is set to a single exposure level.
- The alignment and calibration techniques may be used for initial system set up and initial video streaming. Further, they may be used once the system is operative and is otherwise streaming video. In such a context, real world problems occur, such as camera faults and subsequent camera displacement or offset. Techniques of the invention address these issues.
-
FIG. 5 illustrates processing associated with one such technique. Theimage processor 114 streams video frommultiple cameras 500. Theimage processor 114 is configured to check for a faulty camera 502 (e.g., a lost signal). If a faulty camera exists, theimage processor 114 maximizes the field of view ofadjacent cameras 504. As shown inFIG. 1 , an overlappingregion 110 exists between cameras. So for example, ifcamera 100 is faulty, the full field ofview 108 ofcamera 106 would be utilized. -
FIG. 6 illustrates processing associated with a camera misalignment after initial alignment and calibration. Theimage processor 114 streams video frommultiple cameras 600. Theimage processor 114 is configured to check for an offset camera 602 (e.g., a camera misaligned from its original position). If an offset camera is identified, the offset is evaluated 604. For example, the displacement between matching pixels of adjacent video streams may be evaluated. Also, outside sensors may be used to determine this displacement, such as accelerometers or digital gyroscopes. This displacement is then used to form updated alignment parameters (e.g., the original alignment parameters are compensated to incorporate the identified displacement) 606. Video streams are subsequently stitched based on the updatedalignment parameters 608. - An embodiment of the present invention relates to a computer storage product with a computer readable storage medium having computer code thereon for performing various computer-implemented operations. The media and computer code may be those specially designed and constructed for the purposes of the present invention, or they may be of the kind well known and available to those having skill in the computer software arts. Examples of computer-readable media include, but are not limited to: magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROMs, DVDs and holographic devices; magneto-optical media; and hardware devices that are specially configured to store and execute program code, such as application-specific integrated circuits (“ASICs”), programmable logic devices (“PLDs”) and ROM and RAM devices. Examples of computer code include machine code, such as produced by a compiler, and files containing higher-level code that are executed by a computer using an interpreter. For example, an embodiment of the invention may be implemented using JAVA®, C++, or other object-oriented programming language and development tools. Another embodiment of the invention may be implemented in hardwired circuitry in place of, or in combination with, machine-executable software instructions.
- The foregoing description, for purposes of explanation, used specific nomenclature to provide a thorough understanding of the invention. However, it will be apparent to one skilled in the art that specific details are not required in order to practice the invention. Thus, the foregoing descriptions of specific embodiments of the invention are presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the invention to the precise forms disclosed; obviously, many modifications and variations are possible in view of the above teachings. The embodiments were chosen and described in order to best explain the principles of the invention and its practical applications, they thereby enable others skilled in the art to best utilize the invention and various embodiments with various modifications as are suited to the particular use contemplated. It is intended that the following claims and their equivalents define the scope of the invention.
Claims (12)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/691,632 US20130141526A1 (en) | 2011-12-02 | 2012-11-30 | Apparatus and Method for Video Image Stitching |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201161566269P | 2011-12-02 | 2011-12-02 | |
US13/691,632 US20130141526A1 (en) | 2011-12-02 | 2012-11-30 | Apparatus and Method for Video Image Stitching |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130141526A1 true US20130141526A1 (en) | 2013-06-06 |
Family
ID=48523714
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/691,632 Abandoned US20130141526A1 (en) | 2011-12-02 | 2012-11-30 | Apparatus and Method for Video Image Stitching |
US13/691,654 Active 2033-08-26 US9516225B2 (en) | 2011-12-02 | 2012-11-30 | Apparatus and method for panoramic video hosting |
US15/366,878 Active US9843840B1 (en) | 2011-12-02 | 2016-12-01 | Apparatus and method for panoramic video hosting |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/691,654 Active 2033-08-26 US9516225B2 (en) | 2011-12-02 | 2012-11-30 | Apparatus and method for panoramic video hosting |
US15/366,878 Active US9843840B1 (en) | 2011-12-02 | 2016-12-01 | Apparatus and method for panoramic video hosting |
Country Status (1)
Country | Link |
---|---|
US (3) | US20130141526A1 (en) |
Cited By (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140168357A1 (en) * | 2012-12-13 | 2014-06-19 | Microsoft Corporation | Displacing image on imager in multi-lens cameras |
US20160307300A1 (en) * | 2013-12-06 | 2016-10-20 | Huawei Device Co. Ltd. | Image processing method and apparatus, and terminal |
US9516225B2 (en) | 2011-12-02 | 2016-12-06 | Amazon Technologies, Inc. | Apparatus and method for panoramic video hosting |
CN106447735A (en) * | 2016-10-14 | 2017-02-22 | 安徽协创物联网技术有限公司 | Panoramic camera geometric calibration processing method |
CN106470313A (en) * | 2015-08-23 | 2017-03-01 | 宏达国际电子股份有限公司 | Image generation system and image generating method |
US9723223B1 (en) | 2011-12-02 | 2017-08-01 | Amazon Technologies, Inc. | Apparatus and method for panoramic video hosting with directional audio |
US9781356B1 (en) | 2013-12-16 | 2017-10-03 | Amazon Technologies, Inc. | Panoramic video viewer |
US9826156B1 (en) | 2015-06-16 | 2017-11-21 | Amazon Technologies, Inc. | Determining camera auto-focus settings |
US9838687B1 (en) | 2011-12-02 | 2017-12-05 | Amazon Technologies, Inc. | Apparatus and method for panoramic video hosting with reduced bandwidth streaming |
US20170352191A1 (en) * | 2016-06-07 | 2017-12-07 | Visbit Inc. | Virtual Reality 360-Degree Video Camera System for Live Streaming |
US9854155B1 (en) | 2015-06-16 | 2017-12-26 | Amazon Technologies, Inc. | Determining camera auto-focus settings |
CN107592452A (en) * | 2017-09-05 | 2018-01-16 | 深圳市圆周率软件科技有限责任公司 | A kind of panorama audio-video acquisition equipment and method |
TWI617195B (en) * | 2017-06-22 | 2018-03-01 | 宏碁股份有限公司 | Image capturing apparatus and image stitching method thereof |
US9998661B1 (en) | 2014-05-13 | 2018-06-12 | Amazon Technologies, Inc. | Panoramic camera enclosure |
WO2018171487A1 (en) * | 2017-03-23 | 2018-09-27 | 华为技术有限公司 | Panoramic video playback method and client terminal |
US10432855B1 (en) * | 2016-05-20 | 2019-10-01 | Gopro, Inc. | Systems and methods for determining key frame moments to construct spherical images |
US10958950B2 (en) | 2016-03-22 | 2021-03-23 | Interdigital Vc Holdings, Inc. | Method, apparatus and stream of formatting an immersive video for legacy and immersive rendering devices |
US11025804B2 (en) * | 2019-06-27 | 2021-06-01 | Intel Corporation | Modular wedges for generating panoramic video |
US11305724B2 (en) | 2020-04-30 | 2022-04-19 | Zoox, Inc. | Pedestrian protection system for sensor pods |
US11353590B2 (en) * | 2020-04-30 | 2022-06-07 | Zoox, Inc. | Sensor pod calibration |
US11481026B2 (en) | 2019-08-22 | 2022-10-25 | Samsung Electronics Co., Ltd. | Immersive device and method for streaming of immersive media |
US11558431B2 (en) | 2017-01-05 | 2023-01-17 | Ricoh Company, Ltd. | Communication terminal, communication system, communication method, and display method |
US11623585B2 (en) | 2020-04-30 | 2023-04-11 | Zoox, Inc. | Sensor pod coverage and placement on vehicle |
US11760313B2 (en) | 2020-04-30 | 2023-09-19 | Zoox, Inc. | Sensor pod cleaning system |
US11856297B1 (en) * | 2014-12-31 | 2023-12-26 | Gn Audio A/S | Cylindrical panorama hardware |
US11953623B2 (en) | 2020-04-30 | 2024-04-09 | Zoox, Inc. | Sensor pod assembly |
Families Citing this family (61)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9389677B2 (en) * | 2011-10-24 | 2016-07-12 | Kenleigh C. Hobby | Smart helmet |
WO2013086246A1 (en) | 2011-12-06 | 2013-06-13 | Equisight Inc. | Virtual presence model |
US9836875B2 (en) * | 2013-04-26 | 2017-12-05 | Flipboard, Inc. | Viewing angle image manipulation based on device rotation |
US8917355B1 (en) | 2013-08-29 | 2014-12-23 | Google Inc. | Video stitching system and method |
US8818081B1 (en) | 2013-10-16 | 2014-08-26 | Google Inc. | 3D model updates using crowdsourced video |
US9300882B2 (en) * | 2014-02-27 | 2016-03-29 | Sony Corporation | Device and method for panoramic image processing |
US9699437B2 (en) * | 2014-03-03 | 2017-07-04 | Nextvr Inc. | Methods and apparatus for streaming content |
US10477189B2 (en) | 2014-04-30 | 2019-11-12 | Intel Corporation | System and method of multi-view reconstruction with user-selectable novel views |
KR101530826B1 (en) * | 2014-08-18 | 2015-06-24 | 정관호 | Playing method and the system of 360 degree spacial video |
US9942583B2 (en) | 2014-11-18 | 2018-04-10 | The Invention Science Fund Ii, Llc | Devices, methods and systems for multi-user capable visual imaging arrays |
US9866765B2 (en) * | 2014-11-18 | 2018-01-09 | Elwha, Llc | Devices, methods, and systems for visual imaging arrays |
US10027873B2 (en) | 2014-11-18 | 2018-07-17 | The Invention Science Fund Ii, Llc | Devices, methods and systems for visual imaging arrays |
US10491796B2 (en) | 2014-11-18 | 2019-11-26 | The Invention Science Fund Ii, Llc | Devices, methods and systems for visual imaging arrays |
US10362290B2 (en) * | 2015-02-17 | 2019-07-23 | Nextvr Inc. | Methods and apparatus for processing content based on viewing information and/or communicating content |
KR102533555B1 (en) | 2015-02-17 | 2023-05-18 | 네버마인드 캐피탈 엘엘씨 | Methods and apparatus for generating and using reduced resolution images and/or communicating such images to a playback or content distribution device |
GB2536025B (en) * | 2015-03-05 | 2021-03-03 | Nokia Technologies Oy | Video streaming method |
TW201637432A (en) * | 2015-04-02 | 2016-10-16 | Ultracker Technology Co Ltd | Real-time image stitching device and real-time image stitching method |
US20160353146A1 (en) * | 2015-05-27 | 2016-12-01 | Google Inc. | Method and apparatus to reduce spherical video bandwidth to user headset |
US10582125B1 (en) * | 2015-06-01 | 2020-03-03 | Amazon Technologies, Inc. | Panoramic image generation from video |
US10506006B2 (en) | 2015-09-09 | 2019-12-10 | Vantrix Corporation | Method and system for flow-rate regulation in a content-controlled streaming network |
US10419770B2 (en) | 2015-09-09 | 2019-09-17 | Vantrix Corporation | Method and system for panoramic multimedia streaming |
US10694249B2 (en) | 2015-09-09 | 2020-06-23 | Vantrix Corporation | Method and system for selective content processing based on a panoramic camera and a virtual-reality headset |
US11108670B2 (en) | 2015-09-09 | 2021-08-31 | Vantrix Corporation | Streaming network adapted to content selection |
US11287653B2 (en) | 2015-09-09 | 2022-03-29 | Vantrix Corporation | Method and system for selective content processing based on a panoramic camera and a virtual-reality headset |
US9781342B1 (en) * | 2015-10-22 | 2017-10-03 | Gopro, Inc. | System and method for identifying comment clusters for panoramic content segments |
US20170195561A1 (en) * | 2016-01-05 | 2017-07-06 | 360fly, Inc. | Automated processing of panoramic video content using machine learning techniques |
US10334224B2 (en) * | 2016-02-19 | 2019-06-25 | Alcacruz Inc. | Systems and method for GPU based virtual reality video streaming server |
WO2017205642A1 (en) * | 2016-05-25 | 2017-11-30 | Livit Media Inc. | Methods and systems for live sharing 360-degree video streams on a mobile device |
CN106060513B (en) * | 2016-06-29 | 2017-11-21 | 深圳市优象计算技术有限公司 | A kind of code stream caching method for cylinder three-dimensional panoramic video netcast |
CN106131581A (en) * | 2016-07-12 | 2016-11-16 | 上海摩象网络科技有限公司 | The panoramic video manufacturing technology of mixed image |
CN106101847A (en) | 2016-07-12 | 2016-11-09 | 三星电子(中国)研发中心 | The method and system of panoramic video alternating transmission |
CN106385587B (en) * | 2016-09-14 | 2019-08-02 | 三星电子(中国)研发中心 | Share the method, apparatus and system at virtual reality visual angle |
KR102630681B1 (en) | 2016-10-11 | 2024-01-30 | 삼성전자주식회사 | Display apparatus and method for generating capture image |
KR20180051202A (en) * | 2016-11-08 | 2018-05-16 | 삼성전자주식회사 | Display apparatus and control method thereof |
GB2557175A (en) * | 2016-11-17 | 2018-06-20 | Nokia Technologies Oy | Method for multi-camera device |
US10999602B2 (en) | 2016-12-23 | 2021-05-04 | Apple Inc. | Sphere projected motion estimation/compensation and mode decision |
US11259046B2 (en) | 2017-02-15 | 2022-02-22 | Apple Inc. | Processing of equirectangular object data to compensate for distortion by spherical projections |
US10924747B2 (en) | 2017-02-27 | 2021-02-16 | Apple Inc. | Video coding techniques for multi-view video |
CN108574808B (en) * | 2017-03-08 | 2021-02-12 | 杭州海康威视数字技术股份有限公司 | Image splicing method and device and image processing system |
CN106954093B (en) * | 2017-03-15 | 2020-12-04 | 北京小米移动软件有限公司 | Panoramic video processing method, device and system |
US10587800B2 (en) * | 2017-04-10 | 2020-03-10 | Intel Corporation | Technology to encode 360 degree video content |
EP3404913B1 (en) | 2017-05-16 | 2019-11-06 | Axis AB | A system comprising a video camera and a client device and a method performed by the same |
FR3066672B1 (en) * | 2017-05-19 | 2020-05-22 | Sagemcom Broadband Sas | METHOD FOR COMMUNICATING AN IMMERSIVE VIDEO |
US11093752B2 (en) | 2017-06-02 | 2021-08-17 | Apple Inc. | Object tracking in multi-view video |
US11272160B2 (en) * | 2017-06-15 | 2022-03-08 | Lenovo (Singapore) Pte. Ltd. | Tracking a point of interest in a panoramic video |
CN107277474B (en) * | 2017-06-26 | 2019-06-25 | 深圳看到科技有限公司 | Panorama generation method and generating means |
US20190005709A1 (en) * | 2017-06-30 | 2019-01-03 | Apple Inc. | Techniques for Correction of Visual Artifacts in Multi-View Images |
US10754242B2 (en) | 2017-06-30 | 2020-08-25 | Apple Inc. | Adaptive resolution and projection format in multi-direction video |
CN109214983B (en) * | 2017-06-30 | 2022-12-13 | 宏碁股份有限公司 | Image acquisition device and image splicing method thereof |
US11202117B2 (en) * | 2017-07-03 | 2021-12-14 | Telefonaktiebolaget Lm Ericsson (Publ) | Methods for personalized 360 video delivery |
US10621784B2 (en) * | 2017-09-29 | 2020-04-14 | Sony Interactive Entertainment America Llc | Venue mapping for virtual reality spectating of live events |
CN109698952B (en) * | 2017-10-23 | 2020-09-29 | 腾讯科技(深圳)有限公司 | Panoramic video image playing method and device, storage medium and electronic device |
CN110166764B (en) | 2018-02-14 | 2022-03-01 | 阿里巴巴集团控股有限公司 | Visual angle synchronization method and device in virtual reality VR live broadcast |
US10708494B2 (en) * | 2018-08-13 | 2020-07-07 | At&T Intellectual Property I, L.P. | Methods, systems and devices for adjusting panoramic video content |
EP3618442B1 (en) * | 2018-08-27 | 2020-09-30 | Axis AB | An image capturing device, a method and computer program product for forming an encoded image |
US11032590B2 (en) | 2018-08-31 | 2021-06-08 | At&T Intellectual Property I, L.P. | Methods, devices, and systems for providing panoramic video content to a mobile device from an edge server |
CN111354088B (en) * | 2018-12-21 | 2023-11-10 | 宏达国际电子股份有限公司 | Environment map building method and system |
US20220337805A1 (en) * | 2019-09-13 | 2022-10-20 | Sony Group Corporation | Reproduction device, reproduction method, and recording medium |
US11023041B1 (en) * | 2019-11-07 | 2021-06-01 | Varjo Technologies Oy | System and method for producing images based on gaze direction and field of view |
US11412310B2 (en) * | 2020-05-18 | 2022-08-09 | Qualcomm Incorporated | Performing and evaluating split rendering over 5G networks |
DE102022115806A1 (en) | 2022-06-24 | 2024-01-04 | Valeo Comfort And Driving Assistance | Method and system for providing an image to be displayed by an output device |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5319465A (en) * | 1991-09-20 | 1994-06-07 | Sony Pictures Entertainment, Inc. | Method for generating film quality images on videotape |
US6611241B1 (en) * | 1997-12-02 | 2003-08-26 | Sarnoff Corporation | Modular display system |
US20040030527A1 (en) * | 2002-02-07 | 2004-02-12 | Accu-Sport International, Inc. | Methods, apparatus and computer program products for processing images of a golf ball |
US6788333B1 (en) * | 2000-07-07 | 2004-09-07 | Microsoft Corporation | Panoramic video |
US7324664B1 (en) * | 2003-10-28 | 2008-01-29 | Hewlett-Packard Development Company, L.P. | Method of and system for determining angular orientation of an object |
US20100150458A1 (en) * | 2008-12-12 | 2010-06-17 | International Business Machines Corporation | Generating Cohorts Based on Attributes of Objects Identified Using Video Input |
US8334905B2 (en) * | 2010-05-05 | 2012-12-18 | Cisco Technology, Inc. | Zone, system and failure aware self adjusting IP surveillance cameras |
US8406562B2 (en) * | 2006-08-11 | 2013-03-26 | Geo Semiconductor Inc. | System and method for automated calibration and correction of display geometry and color |
US8687070B2 (en) * | 2009-12-22 | 2014-04-01 | Apple Inc. | Image capture device having tilt and/or perspective correction |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6133944A (en) * | 1995-12-18 | 2000-10-17 | Telcordia Technologies, Inc. | Head mounted displays linked to networked electronic panning cameras |
US6043837A (en) * | 1997-05-08 | 2000-03-28 | Be Here Corporation | Method and apparatus for electronically distributing images from a panoptic camera system |
US6192393B1 (en) | 1998-04-07 | 2001-02-20 | Mgi Software Corporation | Method and system for panorama viewing |
US20020049979A1 (en) | 2000-05-18 | 2002-04-25 | Patrick White | Multiple camera video system which displays selected images |
US20020021353A1 (en) | 2000-06-09 | 2002-02-21 | Denies Mark | Streaming panoramic video |
US7796162B2 (en) | 2000-10-26 | 2010-09-14 | Front Row Technologies, Llc | Providing multiple synchronized camera views for broadcast from a live venue activity to remote viewers |
GB0313866D0 (en) * | 2003-06-14 | 2003-07-23 | Impressive Ideas Ltd | Display system for recorded media |
US8126155B2 (en) | 2003-07-02 | 2012-02-28 | Fuji Xerox Co., Ltd. | Remote audio device management system |
US20050280701A1 (en) | 2004-06-14 | 2005-12-22 | Wardell Patrick J | Method and system for associating positional audio to positional video |
US20070035612A1 (en) | 2005-08-09 | 2007-02-15 | Korneluk Jose E | Method and apparatus to capture and compile information perceivable by multiple handsets regarding a single event |
BRPI0622048B1 (en) | 2006-10-20 | 2018-09-18 | Thomson Licensing | method, device, and system for generating regions of interest in video content |
US8339456B2 (en) | 2008-05-15 | 2012-12-25 | Sri International | Apparatus for intelligent and autonomous video content generation and streaming |
US20100050221A1 (en) | 2008-06-20 | 2010-02-25 | Mccutchen David J | Image Delivery System with Image Quality Varying with Frame Rate |
WO2010052558A2 (en) | 2008-11-05 | 2010-05-14 | Easywalk Capital S.A. | System and method for the precise integration of virtual objects to interactive panoramic walk-through applications |
GB0907870D0 (en) | 2009-05-07 | 2009-06-24 | Univ Catholique Louvain | Systems and methods for the autonomous production of videos from multi-sensored data |
US10440329B2 (en) | 2009-05-22 | 2019-10-08 | Immersive Media Company | Hybrid media viewing application including a region of interest within a wide field of view |
US8605783B2 (en) * | 2009-05-22 | 2013-12-10 | Microsoft Corporation | Composite video generation |
US20120208168A1 (en) | 2010-10-11 | 2012-08-16 | Teachscape, Inc. | Methods and systems relating to coding and/or scoring of observations of and content observed persons performing a task to be evaluated |
US8970666B2 (en) | 2011-09-16 | 2015-03-03 | Disney Enterprises, Inc. | Low scale production system and method |
US20130141526A1 (en) | 2011-12-02 | 2013-06-06 | Stealth HD Corp. | Apparatus and Method for Video Image Stitching |
-
2012
- 2012-11-30 US US13/691,632 patent/US20130141526A1/en not_active Abandoned
- 2012-11-30 US US13/691,654 patent/US9516225B2/en active Active
-
2016
- 2016-12-01 US US15/366,878 patent/US9843840B1/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5319465A (en) * | 1991-09-20 | 1994-06-07 | Sony Pictures Entertainment, Inc. | Method for generating film quality images on videotape |
US6611241B1 (en) * | 1997-12-02 | 2003-08-26 | Sarnoff Corporation | Modular display system |
US6788333B1 (en) * | 2000-07-07 | 2004-09-07 | Microsoft Corporation | Panoramic video |
US20040030527A1 (en) * | 2002-02-07 | 2004-02-12 | Accu-Sport International, Inc. | Methods, apparatus and computer program products for processing images of a golf ball |
US7324664B1 (en) * | 2003-10-28 | 2008-01-29 | Hewlett-Packard Development Company, L.P. | Method of and system for determining angular orientation of an object |
US8406562B2 (en) * | 2006-08-11 | 2013-03-26 | Geo Semiconductor Inc. | System and method for automated calibration and correction of display geometry and color |
US20100150458A1 (en) * | 2008-12-12 | 2010-06-17 | International Business Machines Corporation | Generating Cohorts Based on Attributes of Objects Identified Using Video Input |
US8687070B2 (en) * | 2009-12-22 | 2014-04-01 | Apple Inc. | Image capture device having tilt and/or perspective correction |
US8334905B2 (en) * | 2010-05-05 | 2012-12-18 | Cisco Technology, Inc. | Zone, system and failure aware self adjusting IP surveillance cameras |
Cited By (35)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10349068B1 (en) | 2011-12-02 | 2019-07-09 | Amazon Technologies, Inc. | Apparatus and method for panoramic video hosting with reduced bandwidth streaming |
US9516225B2 (en) | 2011-12-02 | 2016-12-06 | Amazon Technologies, Inc. | Apparatus and method for panoramic video hosting |
US9723223B1 (en) | 2011-12-02 | 2017-08-01 | Amazon Technologies, Inc. | Apparatus and method for panoramic video hosting with directional audio |
US9838687B1 (en) | 2011-12-02 | 2017-12-05 | Amazon Technologies, Inc. | Apparatus and method for panoramic video hosting with reduced bandwidth streaming |
US9843840B1 (en) | 2011-12-02 | 2017-12-12 | Amazon Technologies, Inc. | Apparatus and method for panoramic video hosting |
US9094540B2 (en) * | 2012-12-13 | 2015-07-28 | Microsoft Technology Licensing, Llc | Displacing image on imager in multi-lens cameras |
US20140168357A1 (en) * | 2012-12-13 | 2014-06-19 | Microsoft Corporation | Displacing image on imager in multi-lens cameras |
US20160307300A1 (en) * | 2013-12-06 | 2016-10-20 | Huawei Device Co. Ltd. | Image processing method and apparatus, and terminal |
US9870602B2 (en) * | 2013-12-06 | 2018-01-16 | Huawei Device (Dongguan) Co., Ltd. | Method and apparatus for fusing a first image and a second image |
US10015527B1 (en) | 2013-12-16 | 2018-07-03 | Amazon Technologies, Inc. | Panoramic video distribution and viewing |
US9781356B1 (en) | 2013-12-16 | 2017-10-03 | Amazon Technologies, Inc. | Panoramic video viewer |
US9998661B1 (en) | 2014-05-13 | 2018-06-12 | Amazon Technologies, Inc. | Panoramic camera enclosure |
US11856297B1 (en) * | 2014-12-31 | 2023-12-26 | Gn Audio A/S | Cylindrical panorama hardware |
US9854155B1 (en) | 2015-06-16 | 2017-12-26 | Amazon Technologies, Inc. | Determining camera auto-focus settings |
US9826156B1 (en) | 2015-06-16 | 2017-11-21 | Amazon Technologies, Inc. | Determining camera auto-focus settings |
CN106470313A (en) * | 2015-08-23 | 2017-03-01 | 宏达国际电子股份有限公司 | Image generation system and image generating method |
US10958950B2 (en) | 2016-03-22 | 2021-03-23 | Interdigital Vc Holdings, Inc. | Method, apparatus and stream of formatting an immersive video for legacy and immersive rendering devices |
US10432855B1 (en) * | 2016-05-20 | 2019-10-01 | Gopro, Inc. | Systems and methods for determining key frame moments to construct spherical images |
CN109863754A (en) * | 2016-06-07 | 2019-06-07 | 维斯比特股份有限公司 | For 360 degree of video cameras of virtual reality of fluidisation to be broadcast live |
WO2017214291A1 (en) * | 2016-06-07 | 2017-12-14 | Visbit Inc. | Virtual reality 360-degree video camera system for live streaming |
US10652517B2 (en) * | 2016-06-07 | 2020-05-12 | Visbit Inc. | Virtual reality 360-degree video camera system for live streaming |
US20170352191A1 (en) * | 2016-06-07 | 2017-12-07 | Visbit Inc. | Virtual Reality 360-Degree Video Camera System for Live Streaming |
CN106447735A (en) * | 2016-10-14 | 2017-02-22 | 安徽协创物联网技术有限公司 | Panoramic camera geometric calibration processing method |
US11558431B2 (en) | 2017-01-05 | 2023-01-17 | Ricoh Company, Ltd. | Communication terminal, communication system, communication method, and display method |
WO2018171487A1 (en) * | 2017-03-23 | 2018-09-27 | 华为技术有限公司 | Panoramic video playback method and client terminal |
US11303881B2 (en) | 2017-03-23 | 2022-04-12 | Huawei Technologies Co., Ltd. | Method and client for playing back panoramic video |
TWI617195B (en) * | 2017-06-22 | 2018-03-01 | 宏碁股份有限公司 | Image capturing apparatus and image stitching method thereof |
CN107592452A (en) * | 2017-09-05 | 2018-01-16 | 深圳市圆周率软件科技有限责任公司 | A kind of panorama audio-video acquisition equipment and method |
US11025804B2 (en) * | 2019-06-27 | 2021-06-01 | Intel Corporation | Modular wedges for generating panoramic video |
US11481026B2 (en) | 2019-08-22 | 2022-10-25 | Samsung Electronics Co., Ltd. | Immersive device and method for streaming of immersive media |
US11353590B2 (en) * | 2020-04-30 | 2022-06-07 | Zoox, Inc. | Sensor pod calibration |
US11623585B2 (en) | 2020-04-30 | 2023-04-11 | Zoox, Inc. | Sensor pod coverage and placement on vehicle |
US11760313B2 (en) | 2020-04-30 | 2023-09-19 | Zoox, Inc. | Sensor pod cleaning system |
US11305724B2 (en) | 2020-04-30 | 2022-04-19 | Zoox, Inc. | Pedestrian protection system for sensor pods |
US11953623B2 (en) | 2020-04-30 | 2024-04-09 | Zoox, Inc. | Sensor pod assembly |
Also Published As
Publication number | Publication date |
---|---|
US20130141523A1 (en) | 2013-06-06 |
US9843840B1 (en) | 2017-12-12 |
US9516225B2 (en) | 2016-12-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20130141526A1 (en) | Apparatus and Method for Video Image Stitching | |
CN106875339B (en) | Fisheye image splicing method based on strip-shaped calibration plate | |
US8049786B2 (en) | Method for improving image stitching accuracy with lens distortion correction and device for implementing the same | |
CN108257183B (en) | Camera lens optical axis calibration method and device | |
EP2059046B1 (en) | Method and system for combining videos for display in real-time | |
JP5596972B2 (en) | Control device and control method of imaging apparatus | |
US9600863B2 (en) | Method for combining images | |
US9998659B2 (en) | Method and system for adaptive perspective correction of ultra wide-angle lens images | |
EP3163535A1 (en) | Wide-area image acquisition method and device | |
CN108805807B (en) | Splicing method and system for ring scene images | |
CN106600644B (en) | Parameter correction method and device for panoramic camera | |
US7889234B2 (en) | Automatic calibration for camera lens distortion correction | |
CN110288527B (en) | Panoramic aerial view generation method of vehicle-mounted panoramic camera | |
JP2010009417A (en) | Image processing apparatus, image processing method, program and recording medium | |
KR100790887B1 (en) | Apparatus and method for processing image | |
US20130265468A1 (en) | Camera, distortion correction device and distortion correction method | |
CN111199528A (en) | Fisheye image distortion correction method | |
CN108200360A (en) | A kind of real-time video joining method of more fish eye lens panoramic cameras | |
CN105264877A (en) | Systems and methods for calibration of a 360 degree camera system | |
CN104363385A (en) | Line-oriented hardware implementing method for image fusion | |
CN113259642B (en) | Film visual angle adjusting method and system | |
CN113114975B (en) | Image splicing method and device, electronic equipment and storage medium | |
CN114331835A (en) | Panoramic image splicing method and device based on optimal mapping matrix | |
CN109785390B (en) | Method and device for image correction | |
CN109785225B (en) | Method and device for correcting image |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: STEALTH HD CORP., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BANTA, BILL;DONALDSON, GEOFF;REEL/FRAME:029389/0153 Effective date: 20121126 |
|
AS | Assignment |
Owner name: STEALTHHD, INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BANTA, BILL;DONALDSON, GEOFF;REEL/FRAME:033761/0874 Effective date: 20140917 |
|
AS | Assignment |
Owner name: STEALTHHD LLC, WASHINGTON Free format text: CHANGE OF NAME;ASSIGNOR:STEALTHHD, INC.;REEL/FRAME:034749/0590 Effective date: 20141118 |
|
AS | Assignment |
Owner name: AMAZON TECHNOLOGIES, INC., WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:STEALTHHD LLC;REEL/FRAME:038338/0297 Effective date: 20160222 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |