CN117897951A - 2D digital image capturing system and analog 3D digital image and sequence - Google Patents

2D digital image capturing system and analog 3D digital image and sequence Download PDF

Info

Publication number
CN117897951A
CN117897951A CN202280047753.0A CN202280047753A CN117897951A CN 117897951 A CN117897951 A CN 117897951A CN 202280047753 A CN202280047753 A CN 202280047753A CN 117897951 A CN117897951 A CN 117897951A
Authority
CN
China
Prior art keywords
image
scene
display
digital image
images
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202280047753.0A
Other languages
Chinese (zh)
Inventor
杰瑞·尼姆斯
威廉·M·卡尔斯
塞缪尔·波尔
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sai JiuerBoer
Wei LianMKaersi
Jie RuiNimusi
Original Assignee
Sai JiuerBoer
Wei LianMKaersi
Jie RuiNimusi
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US17/355,906 external-priority patent/US20210321077A1/en
Priority claimed from US17/459,067 external-priority patent/US20210392314A1/en
Priority claimed from US17/511,490 external-priority patent/US20220051427A1/en
Priority claimed from US17/525,246 external-priority patent/US20220078392A1/en
Application filed by Sai JiuerBoer, Wei LianMKaersi, Jie RuiNimusi filed Critical Sai JiuerBoer
Priority claimed from US17/834,023 external-priority patent/US20220385807A1/en
Publication of CN117897951A publication Critical patent/CN117897951A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/243Image signal generators using stereoscopic image cameras using three or more 2D image sensors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/128Adjusting depth or disparity
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0081Depth or disparity estimation from stereoscopic image signals

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

A system for capturing a plurality of two-dimensional digital source images of a scene by a user, a smart device having a memory device for storing instructions; a processor in communication with the memory and configured to execute the instructions; a plurality of digital image capturing devices in communication with the processor, each configured to capture a digital image of the scene, the plurality of digital image capturing devices being positioned in linear series within about the inter-pupillary distance, wherein a first digital image capturing device is centered about a position proximate a first end of the inter-pupillary distance, a second digital image capturing device is centered about a position at a second end of the inter-pupillary distance, and any remaining plurality of digital image capturing devices are evenly spaced between the first digital image capturing device and the second digital image capturing device; a display in communication with the processor, the display configured to display the multi-dimensional digital image and add an audio file thereto.

Description

2D digital image capturing system and analog 3D digital image and sequence
Technical Field
The present disclosure relates to 2D image capture, image processing, and analog display of 3D or multi-dimensional images or image sequences.
Background
The Human Visual System (HVS) relies on two-dimensional images to interpret a three-dimensional field of view. By utilizing a mechanism with HVS we create an image/scene that is compatible with HVS.
A mismatch between the point at which the eye has to converge and the distance to which the eye has to focus has negative consequences when viewing a 3D image. While 3D imagery has proven popular and useful in movies, digital advertising, many other applications may be utilized if a viewer is able to view 3D images without wearing specialized glasses or headphones, which is a well-known problem. Misalignment in these systems can lead to image jerkiness, defocus, or blurred features when viewing digital multi-dimensional images. Viewing these images can result in headache and nausea.
In natural viewing, the images arrive at the eyes with different binocular disparities, so that when a viewer looks from one point to another in a visual scene, they have to adjust the vergence of the eyes. The distance at which the lines of sight intersect is the vergence distance. Failure to converge at this distance can result in a dual image. The viewer also adjusts the power of each eye lens (i.e., accommodation) appropriately for the fixed portion of the scene. The distance to which the eye must focus is the accommodation distance. An image that is blurred may appear if not adjusted to this distance. The convergence and accommodation response are coupled in the brain, in particular, the change in convergence drives the change in accommodation, which drives the change in convergence. This coupling is advantageous in natural viewing, since the convergence distance and the accommodation distance are almost always the same.
In the 3D image, the image has different binocular parallax, thereby stimulating convergence change as occurs in natural viewing. However, the adjustment distance is still fixed at the display distance from the viewer, and therefore, the natural correlation between the convergence distance and the adjustment distance is broken, resulting in a so-called convergence-adjustment conflict. This conflict causes several problems. First, the different parallax and focus information may lead to perceived depth distortion. Second, viewers experience difficulty in fusing (fuse) and focusing on critical subjects in the image at the same time. Finally, attempts to adjust convergence and accommodation, respectively, can result in visual discomfort and fatigue for the viewer.
The perception of depth is based on various cues, binocular disparity (binocular disparity) and motion parallax (motion parallaxes) generally provide more accurate depth information than image cues. Binocular parallax and motion parallax provide two independent quantitative cues for depth perception. Binocular disparity refers to the difference in position of a point between two retinal image projections in 3D space.
Conventional stereoscopic displays force viewers to attempt to decouple these processes because they must keep the adjustment at a fixed distance when they must dynamically change the convergence angle to view objects at different stereoscopic distances, otherwise the entire display will slide out of focus. Such decoupling can create eye strain and compromise image quality when viewing such displays.
Recently, some photographers are shooting multiple frames of a scene with cameras of the 80 th century, such as nimlo and NASHIKA 3d 35mm analog film cameras or digital cameras moving between points, developing multiple frames of film from the analog camera, uploading the images to image software, such as PHOTOSHOP, and arranging the images to create a wobble map, moving GIF effect.
Thus, it is apparent that there is a significant unmet need for an intelligent device with an integrated 2D digital image capture system, image manipulation application, and display of a 3D digital image or image sequence that can be configured to address at least some aspects of the problems discussed above.
Disclosure of Invention
Briefly, in example embodiments, the present disclosure may overcome the above-mentioned drawbacks, and may meet the recognized need for a 2D image capture system and display of 3D digital images and 3D sequences, a smart device having: memory means for storing instructions; a processor in communication with the memory, and configured to execute the instructions; a plurality of digital image capturing devices in communication with the processor and each configured to capture a digital image of the scene, the plurality of digital image capturing devices positioned in linear series within about the inter-pupillary distance width, wherein a first digital image capturing device is centered at a first end proximate the inter-pupillary distance width, a second digital image capturing device is centered at a second end proximate the inter-pupillary distance width, and any remaining plurality of digital image capturing devices are evenly spaced between the first digital image capturing device and the second digital image capturing device, the processing step to configure the data set; and a display configured to display the simulated multi-dimensional digital image sequence and/or multi-dimensional digital image, store the multi-dimensional digital image sequence and/or multi-dimensional digital image via a blockchain, store an audio file for playback when viewing the image file, store an associated verification document that authenticates the image or audio file, transmit the stored file, blockchain store such file, create an irreplaceable asset, irreplaceable token (NFT) of such stored file.
Thus, the system and method of use are characterized by: it has the capability to capture multiple images of a scene with a 2D capture device positioned approximately the intra-ocular or inter-pupillary distance width IPD (distance between pupils of the human visual system).
Thus, a feature of the system and method of use is that it is capable of converting an input 2D source image into a sequence of multi-dimensional/multi-spectral images. The output image follows the rule that the "key subject point" remains within the optimal disparity to maintain a clear and distinct image.
Thus, the system and method of use are characterized by: it has the ability to display an analog multi-dimensional digital image sequence using existing viewing devices.
Thus, the system and method of use are characterized by: it has the ability to take a sequence of multi-dimensional digital images, view the sequence of multi-dimensional digital images, and transmit the sequence of multi-dimensional digital images over the internet. This stand-alone system may be integrated into a smart phone, tablet, or used with an external device. A series of 4 camera shots enables us to produce a special motion parallax image-DIGY, which can be viewed without a special screen. The system may be used in a fully automatic mode or in a manual mode for operator interaction with the scene.
Thus, the system and method of use are characterized by the following capabilities: the viewing device or other viewing function is integrated into the display such as a barrier screen (black lines), lenticular lenses, arcs, curved surfaces, trapezoids, parabolas, overlays, waveguides, black lines, etc. with integrated LCD layers in the LED or OLED, LCD, OLED and combinations thereof or other viewing devices.
Another feature of the digital multidimensional image platform based system and method of use is: the ability to generate digital multi-dimensional images that can be viewed on viewing screens such as mobile and landline phones, smartphones (including iphones), tablet computers, notebook computers, monitors and other displays and/or specific output devices, directly without the need for 3D glasses or headphones.
In an exemplary embodiment, a system for simulating a 3D image sequence from a series of 2D images of a scene includes a smart device having: memory means for storing instructions; a processor in communication with the memory device, the processor configured to execute the instructions; a plurality of digital image capturing devices in communication with the processor and each configured to capture a digital image of the scene, the plurality of digital image capturing devices being positioned approximately within about the inter-pupillary distance width in a linear series, wherein a first digital image capturing device is positioned near a first end of the inter-pupillary distance width, a second digital image capturing device is positioned near a second end of the inter-pupillary distance width, and any remaining plurality of digital image capturing devices are evenly spaced between the first and second digital image capturing devices to capture a series of 2D images of the scene; a display in communication with the processor, the display configured to display the sequence of multi-dimensional digital images and superimpose an audio file over the sequence of multi-dimensional digital images via input on the display.
Features of the present disclosure may include a system having a series of capture devices, such as two, three, four or more, such multiple capture devices (digital image cameras) positioned linearly within the eye or inter-pupillary distance width-distance between pupils of an average person, the system capturing and storing 2D source images of two, three, four or more, multiple scenes, the system marking and identifying the images based on the source capture device capturing the images.
Features of the present disclosure may include a system having a display device configured from a stack of components such as a top glass cover, capacitive touch screen glass, polarizers, diffusers, and backlight. In addition, image sources, such as LCDs, such as LED, ELED, PDP, QLED, and other types of display technologies. In addition, the display device may include a lens array, preferably located between the capacitive touch screen glass and the LCD panel assembly stack, configured to bend or refract light in a manner that enables display of high quality 2D images and left-right interleaved stereoscopic image pairs as 3D or multi-dimensional digital images of the scene.
The characteristics of the present disclosure are: the ability to overcome the above-mentioned drawbacks by another important parameter for determining convergence points or key subject points, because viewing images that are not aligned with key subject points can be confusing to the human visual system, resulting in image blurring and dual images.
The characteristics of the present disclosure are: the ability to select a convergence point or key subject point anywhere between a near or near plane and a far or back plane, manual mode user selection.
The characteristics of the present disclosure are: the ability to overcome the above-mentioned drawbacks by another important parameter for determining the comfort circle CoC, because viewing an image that is not aligned with the comfort circle CoC can be confusing to the human visual system, resulting in image blurring and double images.
The characteristics of the present disclosure are: the ability to overcome the above-mentioned drawbacks by another important parameter for determining a comfort circle CoC that is fused with a point of view arc or point and Panum area is overcome, because viewing an image that is not aligned with a comfort circle CoC that is fused with a point of view arc or point and Panum area can be confusing to the human visual system, resulting in image blurring and double images.
The characteristics of the present disclosure are: with the ability to overcome the above-mentioned drawbacks by determining another important parameter of the gray depth map, the system interpolates intermediate points based on the specified points (closest, key subject, and farthest points) in the scene, the system assigns values to these intermediate points, and renders the sum thereof into the gray depth map. The gray map uses values assigned to different points (closest point, key subject point, and farthest point) in the scene to generate volumetric disparities. This mode also allows for assigning volumetric disparities or fillets to a single object in the scene.
A feature of the present disclosure is its ability to measure the depth or z-axis of an object or object element and/or to compare based on a known size of an object in a scene.
The characteristics of the present disclosure are: it has the ability to manually or automatically select key subjects from a plurality of images of a scene displayed on a display using key subject algorithms and to generate a sequence of multi-dimensional digital images for viewing on the display.
The present disclosure is characterized by its following capabilities: image alignment, horizontal image translation, or editing algorithms are utilized to manually or automatically align multiple images of a scene around a critical subject for display.
The features of the present disclosure are: it utilizes an image translation algorithm to align key subject points of two images of a scene for display.
The characteristics of the present disclosure are: it has the ability to generate DIFYS (differential image format), a specific technique for obtaining multiple views of a scene and creating a series of images that produce depth without glasses or any other viewing aids. The system creates a 3D view using horizontal image translation in conjunction with a form of motion parallax. DIFYS is created by the observer's eyes flipping different views of a single scene. These views are captured by the motion of an image capture system or by scenes taken by multiple cameras, with each camera in the array looking at a different location.
According to a first aspect of the present disclosure, simulating a 3D image sequence from a 2D image frame sequence may be used to capture a plurality of 2D image frames (images) of a scene from a plurality of different viewpoints, wherein a first proximal plane and a second distal plane are identified within each image frame in the sequence, and wherein each viewpoint substantially maintains the same first proximal image plane of each image frame; determining a depth estimate for a first near side plane and a second far side plane within each image frame in the sequence, aligning the first near side plane of each image frame in the sequence, and shifting the second far side plane of each subsequent image frame in the sequence based on the depth estimate for the second far side plane of each image frame to produce a modified image frame corresponding to each 2D image frame and displaying the modified image frames in sequence.
The present disclosure changes the focus of objects at different planes in a display scene to match the requirements of vergence and stereoscopic retinal disparity to better simulate natural viewing conditions. By adjusting the focus of key objects in the scene to match their stereo retinal disparity, cues for eyeball accommodation and convergence are consistent. As in natural vision, viewers focus on different objects by changing accommodations. Natural viewing conditions are better simulated and eye fatigue is reduced due to the reduced mismatch between accommodation and convergence.
The present disclosure may be used to determine three or more planes for each image frame in a sequence.
Furthermore, it is preferred that the planes have different depth estimates.
Furthermore, it is preferred that each respective plane is shifted based on a difference between the depth estimate of the respective plane and the first proximal plane.
Preferably, the first proximal plane of each modified image frame is aligned such that the first proximal planes are positioned in the same pixel space.
It is further preferred that the first plane comprises key principal points.
Preferably, the planes comprise at least one foreground plane.
Furthermore, it is preferred that the planes comprise at least one background plane.
Preferably, the successive observation points lie on a straight line.
According to a second aspect of the present invention there is provided a non-transitory computer readable storage medium storing instructions which, when executed by a processor, cause the processor to perform a method according to the second aspect of the present invention.
These and other features of the smart device with a 2D digital image capture system, an image manipulation application, and display of a 3D digital image or sequence of images will become more apparent to those skilled in the art from the foregoing abstract and the following description of the drawings, detailed description, and claims, read in light of the accompanying drawings or illustrations.
Drawings
The disclosure will be better understood by reading the detailed description of the preferred and selected alternative embodiments with reference to the drawings in which like numerals indicate like structure and in which like elements are referenced throughout, and in which:
FIG. 1A illustrates a 2D rendering of an image based on a change in the direction of a viewer relative to a display;
FIG. 1B illustrates a 2D rendering of an image with binocular disparity (disparity) due to horizontal split disparity (parallaxes) for the left and right eyes;
FIG. 2A is an illustration of a cross-sectional view of a human eye structure;
FIG. 2B is a graph of density versus fovea for rod and cone cells;
FIG. 3 is a top view of an observer's field of view;
FIG. 4A is a side view illustration identifying a plane of a scene captured using a camera or other capturing device;
fig. 4B A is a front view of an exemplary embodiment of two images of the scene in fig. 4A captured with the capture device shown in fig. 8G;
FIG. 5 is a top view of a comfort circle identifying the plane of the scene and to the scale of FIG. 4;
FIG. 6 is a block diagram of a computer system of the present disclosure;
FIG. 7 is a block diagram of a communication system implemented by the computer system of FIG. 1;
FIG. 8A is a diagram of an exemplary embodiment of a computing device with four image capture devices positioned in vertical linear series within an intra-ocular or inter-pupillary distance (i.e., distance between pupils of an average person);
FIG. 8B is a diagram of an exemplary embodiment of a computing device with four image capture device levels positioned in linear series within the intra-ocular or inter-pupillary distance (i.e., the distance between pupils of an average person);
FIG. 8C is an exploded view of an exemplary embodiment of the four image capture devices of FIGS. 8A and 8B in linear series;
FIG. 8D is a cross-sectional view of an exemplary embodiment of the four image capture devices of FIGS. 8A and 8B in linear series;
fig. 8E is an exploded view of an exemplary embodiment of three image capturing devices in linear series within an intra-ocular or inter-pupillary distance (i.e., distance between pupils of an average person);
FIG. 8F is a cross-sectional view of an exemplary embodiment of the three image capture devices of FIG. 8E in linear series;
FIG. 8G is an exploded view of an exemplary embodiment of two image capture devices in linear series within an intra-ocular or inter-pupillary distance (i.e., distance between pupils of an average person);
FIG. 8H is a cross-sectional view of an exemplary embodiment of two image capture devices in linear series of FIG. 8G;
Fig. 9 is a diagram of an exemplary embodiment of a human eye's inter-ocular or intra-pupillary distance width (i.e., the distance between the pupils of an average person);
FIG. 10 is a top view of a plane and comfort circle identifying a scaled scene, wherein a right triangle defines the positioning of the capture device on the lens plane;
FIG. 10A is a top view of an exemplary embodiment identifying right triangles for calculating the comfort circle radius of FIG. 10;
FIG. 10B is a top view of an exemplary embodiment identifying a right triangle for calculating the linear positioning of the capture device on the lens plane of FIG. 10;
FIG. 10C is a top view of an exemplary embodiment of a right triangle identifying the optimal distance for computing the back plate of FIG. 10;
FIG. 11 is a diagram of an exemplary embodiment such as a geometric shift of a point between two images (frames) in FIG. 11A in accordance with selected embodiments of the present disclosure;
FIG. 11A is a front top view of an exemplary embodiment of four images of a scene captured with the capture device shown in FIGS. 8A-8F and aligned with respect to a critical subject point;
FIG. 11B is a front view of an exemplary embodiment of four images of a scene captured with the capture device shown in FIGS. 8A-8F and aligned with respect to a critical subject point;
FIG. 12 is an exemplary embodiment of a flowchart of a method of generating a sequence of multi-dimensional images captured using the capture device shown in FIGS. 8A-8H;
FIG. 13 is an exemplary embodiment of a display with user interactive content for selecting a photography option for a computer system;
FIG. 14A is a top view identifying two frames captured with the capture device shown in FIGS. 8A-8F, showing a near-planar object offset between the key subject and the two frames aligned as shown in FIG. 11B;
FIG. 14B is a top view of an exemplary embodiment of left and right eye virtual depths offset by an object between two frames of FIG. 14A;
FIG. 15A is a cross-sectional view of an exemplary embodiment of a display stack according to selected embodiments of the present disclosure;
FIG. 15B is a cross-sectional view of an exemplary embodiment of an arcuate or curved lens tracking RGB light therethrough according to selected embodiments of the present disclosure;
FIG. 15C is a cross-sectional view of a prototype embodiment of a trapezoidal shaped lens according to a selected embodiment of the present disclosure that tracks RGB light therethrough; |
FIG. 15D is a cross-sectional view of an exemplary embodiment of a dome lens tracking RGB light therethrough according to selected embodiments of the present disclosure;
Fig. 16A is an illustration of an exemplary embodiment of inter-pixel phase processing (pixel interphase processing) of an image (frame) such as in fig. 8A, in accordance with selected embodiments of the present disclosure;
FIG. 16B is a top view of an exemplary embodiment of a display of a computer system running an application;
FIG. 17 is a top view of an exemplary embodiment of viewing a multi-dimensional digital image on a display, wherein the image is within a comfort circle, near a point of co-sight arc or point, within a Panum region, and viewed from a viewing distance;
fig. 18 is an exemplary embodiment of a flowchart of a method of generating a non-replaceable token (NFT) using a computing device having the image capture device shown in fig. 8A-8H;
FIG. 19 is an exemplary embodiment of a flowchart of a method of selecting DIGY sequences and adding audio files thereto using a computing device having the image capture device shown in FIGS. 8A-8H; a kind of electronic device
Fig. 20 is an exemplary embodiment of a flowchart of a method of generating DIGY and adding an audio file thereto using a computing device having the image capture device shown in fig. 8A-8H.
It is to be noted that the drawings are presented for purposes of illustration only and, therefore, are not intended to be limiting of the disclosure to any or all of the exact construction details shown, unless they may be deemed critical to the claimed disclosure.
Detailed Description
In describing exemplary embodiments of the present disclosure, as illustrated, specific terminology is employed for the sake of clarity. However, the disclosure is not intended to be limited to the specific terminology so selected, and it is to be understood that each specific element includes all technical equivalents that operate in a similar manner to accomplish a similar function. The claimed invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. The examples listed herein are non-limiting examples and are merely examples among other possible examples.
The perception of depth is based on various cues, binocular disparity (binocular disparity) and motion parallax (motion parallaxes) generally provide more accurate depth information than image cues. Binocular parallax and motion parallax provide two independent quantitative cues for depth perception. Binocular disparity refers to the difference in position of a point between two retinal image projections in 3D space. As shown in fig. 1A and 1B, the concept of depth of Lu Jian obtained when viewing an object 102 in an image scene 110 shows that: the brain may calculate depth from binocular disparity cues only. In binocular vision, binocular single vision 112 is the location of a point in space that has the same parallax as fixed point 114. An object located on a horizontal line passing through the fixed point 114 will produce a single image, while an object located some distance from this line will produce two images 116, 118.
Classical motion parallax depends on both eye functions. One is the eye's tracking of motion (eye movement to fix motion at a single point) and the second is smooth motion disparity, resulting in parallax or binocular parallax. Classical motion parallax refers to the case where the observer is stationary and the scene surrounding the observer is panning, or vice versa.
By using two images 116, 118 of the same object 102 obtained from slightly different angles, the distance to the object 102 can be triangulated with high accuracy. The angles at which each eye views the object 102 seen by the left eye 104 and the right eye 106 are slightly different. This occurs because of the horizontal separation parallax of the eyes. If an object is far away, the parallax 108 of the image 110 falling on both retinas will be small. If the object is very close or very close, the parallax 108 of the image 110 falling on both retinas will be very large.
Motion parallax 120 refers to the relative image motion (between objects at different depths) due to translation of observer 104. Motion parallax 120 may also provide accurate depth perception, separate from binocular depth cues and picture depth cues, provided that it is accompanied by an auxiliary signal specifying a change in eye direction relative to visual scene 110. As shown, as the eye direction 104 changes, the apparent relative movement of the object 102 with respect to the background gives a cue as to its relative distance. If the object 102 is far away, the object 102 appears to be stationary. If the object 102 is very close or in close proximity, the object 102 appears to move faster.
To see the object 102 in close proximity and fuse the images on the two retinas into one object, the optical axes of the two eyes 104, 106 converge on the object 102. The muscular action of changing the focal length of the eye's lens so as to place a focused image in the fovea of the retina is called accommodation. Both muscle action and unfocusing at adjacent depths provide the brain with additional information that can be used to perceive depth. Image sharpness is a blurred depth cue. However, by changing the focal plane (which appears to be closer and/or farther than the object 102), the ambiguity problem can be resolved.
Fig. 2A and 2B show graphical representations of the anatomy of an eye 200 and the distribution of rod cells and cone cells, respectively. The fovea 202 is responsible for acute foveal vision (also referred to as foveal vision), which is necessary where visual detail is of paramount importance. Fovea 202 is a depression of retinal inner surface 205, about 1.5mm wide, and is entirely comprised of cone cells 204 dedicated to maximum vision. Rod cells 206 are low intensity receptors that receive gray scale information and are important for peripheral vision, while cone cells 204 are high intensity receptors that receive color vision information. The importance of fovea 202 can be more clearly understood with reference to fig. 2B, which shows the distribution of cone cells 204 and rod cells 206 in eye 200. As shown, most of the cone cells 204 that provide the highest visual acuity are located within an angle of 1.5 ° around the center of the fovea 202.
The importance of fovea 202 can be more clearly understood with reference to fig. 2B, which shows the distribution of cone cells 204 and rod cells 206 in eye 200. As shown, most of the cone cells 204 that provide the highest visual acuity lie within an angle of 1.5 ° around the center of the fovea 202.
Fig. 3 illustrates a typical field of view 300 of the Human Visual System (HVS). As shown, the fovea 202 sees only the center 1.5 (degrees) of the field of view 302, with the preferred field of view 304 being within + -15 (degrees) of the center of the fovea 202. Thus, focusing the object on the fovea depends on the linear dimensions, viewing angle, and viewing distance of the object 102. A large object 102 viewed at near will have a large viewing angle outside the foveal vision, while a small object 102 viewed at far will have a small viewing angle inside the foveal vision. Objects 102 falling within foveal vision will produce a high visual acuity in the mind. However, under natural viewing conditions, the viewer perceives not only passively. Instead, they dynamically scan the visual scene 110 by shifting the fixation point and focus of the eye between objects at different viewing distances. In so doing, accommodation and convergence (the angle between the line of sight of the left eye 104 and the line of sight of the right eye 106) during eye movement must be shifted synchronously to place a new object in clear focus at the center of each retina. Thus, accommodation and convergence are naturally linked together reflectively, such that a change in one process automatically drives a corresponding change in the other process.
Fig. 4A shows a typical view of a scene S to be captured by a camera or digital image capturing device, such as image capturing module 830. The scene S may include four planes defined as: the lens frame is defined as (1) a plane through an image lens or sensor in a recording device or camera (image capture module 830), (2) the key subject plane KSP may be a plane through the sensor focus in the scene (here the pair in the scene), (3) the near plane NP may be a plane through the focus closest to the lens plane (brush B in the foreground), and (4) the far plane FP is a plane through the most focal point (tree T in the background). The relative distances from the image capture module 830 are denoted by N, ks, B. The depth of field of the scene S is defined by the distance between the near plane NP and the far plane FP.
As described above, the sense of depth of a stereoscopic image varies according to the distance between the camera and the key subject, which is referred to as the image capturing distance or KS. The sense of depth is also controlled by the convergence angle and the intraocular distance between the camera capturing each successive image, which affects binocular parallax.
In photography, a "circle of confusion" defines the area of focus captured in the scene S. Thus, the near plane NP, the critical body plane KSP, and the far plane FP are all focused. The area outside this circle is blurred.
Fig. 4B illustrates a typical view of a scene S to be captured by a camera or digital image capturing device, such as the image capturing module 830, and more particularly the image capturing module 830 illustrated in fig. 8G. Two image capturing devices 831 and 832 or any other selected pair 831, 832, 833, 834 may be used to capture multiple digital images of a scene S as a left image 810L and a right image 810R of the scene S, as shown in fig. 8A (multiple digital images). Alternatively, the computer system 10, through the image manipulation application and display 208, may be configured to enable the user U to select or identify two of the image capturing devices 831 (1), 832 (2), 833 (3), or 834 (4) to capture two digital images of the scene S as a left image 810L and a right image 810R of the scene S. The user U may click on or otherwise interact with the selection box 812 to select or identify key subjects KS in the source, left, and right images 810L, 810R of the scene S, as shown in fig. 4B.
Fig. 5 shows the comfort circle (CoC) in proportion to fig. 4.1 and 3.1. A comfort circle (CoC) is defined as a circle formed by passing the diameter of the circle (with the width determined radially by 30 degrees of fig. 3) from the center point on the lens plane-image capture module 830-along the perpendicular line of the key body plane KSP (on the scale of fig. 4). (R is the radius of the comfort circle (CoC)).
Conventional stereoscopic displays force viewers to attempt to decouple these processes because they must keep the adjustment at a fixed distance when they must dynamically change the convergence angle to view objects at different stereoscopic distances, otherwise the entire display will slide out of focus. Such decoupling can create eye strain and compromise image quality when viewing such displays.
Certain variables need to be defined in order to understand the present disclosure. The object field refers to the entire image being constructed. "critical subject point" is defined as the point where the scene converges, i.e., the point in depth of field that remains in focus at all times and has no parallax at the critical subject point. The foreground and background points are the closest and farthest points, respectively, from the viewer. Depth of field is the depth or distance (distance of the depicted foreground to background) created within the object field. The principal axis is a line perpendicular to the scene through the key principal points. Parallax or binocular parallax refers to the difference in position of any point in the first and last images after alignment of a critical subject. In digital patterning, the shift of key principal points from principal axis between frames is always kept as an integer number of pixels from principal axis. The total disparity is the sum of the absolute value of the shift of the key principal point of the nearest frame and the principal axis and the absolute value of the shift of the key principal point of the farthest frame and the principal axis.
Here, the applicant refers to depth of field or circle of confusion when capturing an image, and to a comfort circle when viewing an image on a viewing device.
U.S. patent 9,992,473, U.S. patent 10,033,990, and U.S. patent 10,178,247 are incorporated by reference in their entirety.
It is known to create depth perception using motion parallax. However, in order to maximize depth while maintaining a pleasant viewing experience, a systematic approach was introduced. The system combines factors of the human visual system with the image capture program to produce a realistic depth experience on any 2D viewing device.
This technique introduces a comfort circle (CoC) that specifies the position of the image capture system relative to the scene S. The comfort circle (CoC) sets the best near plane NP and far plane FP with respect to the key subject KS (convergence point, focus), i.e. controls the parallax of the scene S.
The development of this system allows any capturing device, such as an iPhone, camera or video camera, to be used to capture a scene. Similarly, the captured images may be combined and viewed on any digital output device such as a smart phone, tablet, monitor, television, notebook, or computer screen.
As will be appreciated by one of skill in the art, the present disclosure may be embodied as a method, data processing system, or computer program product. Accordingly, the present disclosure may take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present disclosure may take the form of a computer program product on a computer-readable storage medium having computer-readable program code means embodied in the medium. Any suitable computer readable medium may be utilized including hard disks, ROM, RAM, CD-ROMs, electronic, optical, magnetic storage devices, and the like.
The present disclosure is described below with reference to flowchart illustrations of methods, apparatus (systems) and computer program products according to embodiments of the disclosure. It will be understood that each block or step of the flowchart illustrations, and combinations of blocks or steps in the flowchart illustrations, can be implemented by computer program instructions or operations. These computer program instructions may be loaded onto a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions which execute on the computer or other programmable data processing apparatus create means for implementing the functions specified in the flowchart block (s)/step(s).
These computer program instructions or operations may also be stored in a computer-usable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions or operations stored in the computer-usable memory produce an article of manufacture including instruction means that implement the function specified in the flowchart block (s)/step(s). The computer program instructions or operations may also be loaded onto a computer or other programmable data processing apparatus (processor) to cause a series of operational steps to be performed on the computer or other programmable apparatus (processor) to produce a computer implemented process such that the instructions or operations which execute on the computer or other programmable apparatus (processor) provide steps for implementing the functions specified in the flowchart block (s)/step(s).
Accordingly, blocks or steps of the flowchart illustrations support combinations of means for performing the specified functions, combinations of steps for performing the specified functions and program instruction means for performing the specified functions. It will also be understood that each block or step of the flowchart illustrations, and combinations of blocks or steps in the flowchart illustrations, can be implemented by special purpose hardware-based computer systems which perform the specified functions or steps, or combinations of special purpose hardware and computer instructions or operations.
Computer programming for practicing the present disclosure can be written in a variety of programming languages, database languages, and the like. However, it is to be appreciated that other source or object-oriented programming languages, as well as other conventional programming languages, may be utilized without departing from the spirit and intent of the present disclosure.
Referring now to FIG. 6, FIG. 6 illustrates a block diagram of a computer system 10, the computer system 10 providing a suitable environment for implementing embodiments of the present disclosure. The computer architecture shown in fig. 6 is split into two parts-a motherboard 600 and an input/output (I/O) device 620. Motherboard 600 preferably includes interconnections through bus 10: a subsystem or processor for executing instructions, such as a Central Processing Unit (CPU) 602; memory devices such as Random Access Memory (RAM) 604; an input/output (I/O) controller 608; and memory devices, such as Read Only Memory (ROM) 606, also known as firmware. A basic input/output system (BIOS), containing the basic routines that help to transfer information between elements within the computer subsystem, is preferably stored in ROM 606 or is otherwise operably disposed in RAM 604. The computer system 10 also preferably includes: I/O devices 620 such as a main memory device 634 for storing an operating system 626 and executing instructions via applications 624; and a display 628 for visual output; and other suitable I/O devices 632. The main memory 634 is preferably coupled to the CPU 602 through a main memory controller (shown as 608) coupled to the bus 610. The network adapter 630 allows the computer system to send and receive data through the communications device or any other network adapter capable of transmitting and receiving data over a communications link, which may be a wired, optical, or wireless data path. It is recognized herein that Central Processing Unit (CPU) 602 executes instructions, operations or commands stored in ROM 606 or RAM 604.
It is contemplated herein that computer system 10 may include smart devices such as smartphones, iphones, android phones (***, samsung, or other manufacturers), tablet computers, desktop computers, notebook computers, digital image capture devices, and other computing devices (smart devices) having two or more digital image capture devices and/or 3D displays 608.
It is further contemplated herein that the display 608 may be configured as a foldable display or a multi-foldable display that is capable of being unfolded into a larger display surface area.
Many other devices or subsystems or other I/O devices 632 may be connected in a similar manner, including but not limited to devices such as microphones, speakers, flash drives, CD-ROM players, DVD players, printers, main storage devices 634, such as hard disks, and/or modems, etc., each of which is connected through an I/O adapter. In addition, while preferred, not all of the devices shown in FIG. 6 need be present to practice the present disclosure, as discussed below. Further, the devices and subsystems may be interconnected in different configurations than shown in FIG. 6, or may be interconnected based on an optical or gate array, or some combination of elements capable of responding to and performing instructions or operations. The operation of a computer system as shown in fig. 6 is readily known in the art and is not discussed in further detail in this application so as not to unduly complicate this discussion.
Referring now to fig. 7, fig. 2 illustrates a diagram depicting an exemplary communication system 700 in which concepts consistent with the present disclosure may be implemented. An example of each element within the communication system 700 of fig. 7 is described extensively above with reference to fig. 6. In particular, server system 760 and user system 720 have similar attributes to computer system 10 of FIG. 6 and illustrate one possible implementation of computer system 10. The communication system 700 preferably includes one or more user systems 720, 722, 724 (it is contemplated herein that the computer system 10 may include smart devices such as smartphones, iphones, android handsets (***, samsung, or other manufacturers), tablet computers, desktop computers, notebook computers, cameras, and other computing devices (smart devices) with a display 628), one or more server systems 760, and a network 750, which may be, for example, the internet, a public network, a private network, or the cloud. User systems 720 through 724 each preferably include a computer-readable medium, such as random access memory 604, 606, coupled to a processor. The processor-CPU 702-executes program instructions or operations (application software 624) stored in the memory 604, 606. Communication system 700 typically includes one or more user systems 720. For example, user system 720 may include one or more general-purpose computers (e.g., personal computers), one or more special-purpose computers (e.g., devices specifically programmed to communicate with each other and/or server system 760), workstations, servers, devices, digital assistants or "smart" mobile phones or pagers, digital cameras, components, other devices, or some combination of these elements that are capable of responding to and performing instructions or operations.
Similar to the user system 720, the server system 760 preferably includes a computer-readable medium, such as random access memory 604, 606, coupled to a processor. The processor executes program instructions stored in the memory 604, 606. The server system 760 can also include many additional external or internal devices, such as, but not limited to, a mouse, CD-ROM, keyboard, display, storage device, and other attributes similar to the computer system 10 of FIG. 6. Server system 760 may additionally include secondary storage elements, such as database 770 for storing data and information. Although depicted as a single computer system, server system 760 may be implemented as a network of computer processors. The memory 604, 606 in the server system 760 contains one or more executable steps, programs, algorithms, or applications 624 (shown in fig. 6). For example, the server system 760 can include a web server, an information server, an application server, one or more general-purpose computers (e.g., personal computers), one or more special-purpose computers (e.g., devices specifically programmed to communicate with each other), workstations or other devices, or some combination of these elements capable of responding to and executing instructions or operations.
Communication system 700 is capable of communicating and exchanging data (including three-dimensional 3D image files) between user system 720 and server system 760 over communication link 740 and/or network 750. Through the user system 720, the user can communicate data with other user systems 720, 722, 724, and with other systems and devices, such as a server system 760, preferably over a network 750, to electronically transmit, store, print, and/or view multi-dimensional digital primary images. Communication link 740 typically includes a network 750 that communicates directly or indirectly between user systems 720 and server systems 760, regardless of physical separation. Examples of network 750 include the internet, cloud, analog or digital wired and wireless networks, radio, television, cable, satellite, and/or any other delivery mechanism for carrying and/or transmitting data or other information, such as electronically transmitting, storing, printing, and/or viewing a multi-dimensional digital primary image. The communication link 740 may include, for example, a wired, wireless, cable, optical or satellite communication system, or other pathway.
Referring again to fig. 2A, 5 and 8A-8F, for optimal effect and simplified mathematics, the intraocular distance between successive images or frames of the captured scene S is fixed to match the average separation of the left and right eyes of a human to maintain a constant binocular disparity. In addition, the distance from the key subject KS is selected so that the size of the captured key subject image falls within the foveal vision of the observer to produce a high visual acuity of the key subject and maintain a preferred angle of convergence of the viewing angle equal to or less than 15 degrees (15 °).
Fig. 8A-8F disclose an image or frame capturing system for capturing a stereoscopic image (e.g., 2D frames of a 3D sequence) of a scene S, such as fig. 4. Here, when capturing a stereoscopic image of the scene S (e.g., 2D frames of a 3D sequence), the image capture distance, the distance of the image capture system from a point or plane in the scene S-such as the focal length (i.e., zoom in and out) of the key subject KS and camera-may desirably remain unchanged; however, if the spacing between the capturing devices of each successive stereoscopic image is kept constant, the convergence angle will vary accordingly.
Referring now to FIG. 8A, by way of example and not limitation, a computer system 10, such as a smart device or portable smart device, is illustrated having: a back side 810, a first edge such as a short edge 811, and a second edge such as a long edge 812. The backside 810 may include an I/O device 632, such as an exemplary embodiment of an image capture module 830, and may include one or more sensors 840 to measure a distance between the computer system 10 and a selected depth (depth) in an image or scene S. The image capture module 830 may include a plurality of or four digital image capture devices 831, 832, 833, 834 positioned in vertical linear series with respect to the backside 810 or near and parallel to the long edge 812 (distance between pupils of the human visual system in an intra-ocular or inter-pupillary distance width IPD (to optimize the digital multi-dimensional image of the human visual system) the inter-pupillary distance width IPD is preferably the distance between pupils of an average person, may have a distance between about two inches and a half-2.5 inches (6.35 cm), more preferably between about 40mm and 80mm, with the vast majority of adult IPDs being in the range of 50mm to 75mm, a wide range of 45mm to 80mm possibly including (almost) all adults, and the smallest IPD of children (as low as 5 years) being about 40 mm). It is contemplated herein that the plurality of image capturing modules 830 and may include one or more sensors 840 may be configured as a combination of the image capturing device 830 and the sensors 840, configured as an integrated unit or module in which the sensors 840 control the depth of the image capturing device 830 or set the depth of the image capturing device 830, regardless of the different depths in the scene S shown in fig. 4, such as the foreground, and the person P or object, the background (such as the closest point CP, the critical subject point KS, and the furthest point FP). For reference herein, the plurality of image capturing devices may include: a first digital image capturing device 831 centered near the position of the first end ipd.1 of the inter-pupillary distance width IPD, a fourth digital image capturing device 834 centered near the position of the second end ipd.2 of the inter-pupillary distance width IPD, and the remainder digital image capturing devices thereof—second digital image capturing device 832 and third digital image capturing device 833, the second digital image capturing device 832 and third digital image capturing device 833 being uniformly spaced between the first end ipd.1 and the second end ipd.2 of the inter-pupillary distance width IPD, respectively.
It is contemplated herein that a smart device or portable smart device having a display may be configured as rectangular or square or other similar configuration that provides a surface area having a first edge 811 and a second edge 812.
It is contemplated herein that the digital image capturing devices 831 through 834 or the image capturing module 830 may be surrounded by a recessed, stepped, or beveled edge 814, each image capturing device 831 through 834 may be surrounded by a recessed, stepped, or beveled ring 816, and the digital image capturing devices 831 through 834 or the image capturing module 830 may be covered by a lens cover 820 with the lens 818 below.
It is contemplated herein that the digital image capturing devices 831 through 834 may be separate capturing devices rather than part of an image capturing module.
It is further contemplated herein that the digital image capturing devices 831 through 834 may be disposed anywhere on the back side 810 and generally parallel to the long edge 812.
It is contemplated herein that the image capturing means may comprise additional capturing means positioned within the intra-ocular or inter-pupillary distance width IPD.
It is further contemplated herein that the digital image capturing devices 831 through 834 may be used to capture a series of 2D images of the scene S.
Referring now to FIG. 8B, by way of example and not limitation, a computer system 10 or other smart device or portable smart device is shown having a back side 810, a short edge 811, and a long edge 812. The backside 810 may include an I/O device 632, such as an exemplary embodiment of an image capture module 830, and may include one or more sensors 840 to measure a distance between the computer system 10 and a selected depth (depth) in an image or scene S. The image capturing module 830 may comprise a plurality of or four digital image capturing devices 831, 832, 833, 834 positioned in vertical linear series with respect to the backside 810 or near and parallel to the short edge 812 (distance between pupils of the human visual system within a comfortable circular relationship to optimize the digital multi-dimensional image of the human visual system) it is contemplated herein that the plurality of image capturing modules 830 may comprise one or more sensors 840, may be configured as a combination of the image capturing devices 830 and the sensors 840, be configured as an integrated unit or module wherein the sensors 840 control the depth of the image capturing devices 830 or set the depth of the image capturing devices 830, such as different depths in the scene S-such as foreground, background and person P or objects-such as closest point CP, critical subject point KS and far point FP, as shown in fig. 4-the first digital image capturing device 3, the second digital image capturing device 3 at a position near the center of the distance width IPD and the second digital image capturing device at a second end IPD 1, the second digital image capturing device at a distance 3, the second digital image capturing device at the second end IPD 1, and the second digital image capturing device at the second end IPD 3, and the third digital image capturing device at the second end, the same distance 3, respectively.
It is contemplated herein that the digital image capturing devices 831 through 834 or the image capturing module 830 may be surrounded by a recessed, stepped, or beveled edge 814, each image capturing device 831 through 834 may be surrounded by a recessed, stepped, or beveled ring 816, and the digital image capturing devices 831 through 834 or the image capturing module 830 may be covered by a lens cover 820 with the lens 818 below.
It is contemplated herein that the digital image capturing devices 831 through 834 may be separate capturing devices rather than part of an image capturing module.
It is further contemplated herein that the digital image capturing devices 831 through 834 may be disposed anywhere on the back side 810 and generally parallel to the long edge 812.
It is further contemplated herein that the digital image capturing devices 831 through 834 may be used to capture a series of 2D images of the scene S.
With respect to the computer system 10 and the image capture device 830, it should be appreciated that the optimal dimensional relationships, including variations in size, material, shape, form, position, connection, function, and manner of operation, assembly, and use, are intended to be covered by this disclosure.
In the present disclosure, the inter-pupillary distance width IPD may have a measured width to position the digital image capturing devices 831 to 334 center-to-center between approximately a maximum width of 115 millimeters to a minimum width of 50 millimeters; more preferably between a maximum width of about 72.5 millimeters and a minimum width of 53.5 millimeters; and most preferably between a maximum average width of about 64 millimeters and a minimum average width of 61.7 millimeters, and an average width of 63 millimeters (2.48 inches) center-to-center width of the human visual system shown in fig. 9.
Referring again to fig. 1A, 1B, 2A, 5, 9, 14B, binocular parallax is a stereoscopic perception factor that results from left and right eyes equally dividing by about 64 millimeters. When binocular parallax is relatively large, the observer may feel a relatively close distance to the critical subject. When binocular parallax is relatively small, the observer may feel a relatively large distance or greater from the key subject KS. The convergence angle V refers to the angle between the left and right eyes at the apex of the key body when the eyes are focused on the key body KS. As the convergence angle increases (as the eyes rotate inward), the observer considers the distance of the key subject KS relatively small. As the convergence angle decreases (as the eyes rotate outward), the observer considers the distance of the key subject KS relatively large.
Referring now to FIG. 8C, by way of example and not limitation, an exploded view of an exemplary embodiment of an image capture module 830 is shown. The image capturing module 830 may include digital image capturing devices 831 to 834, in which four image capturing devices are linearly connected in series within an intra-eye or inter-pupillary distance width IPD (distance between pupils of an average person). The digital image capturing devices 831 to 834 may include a first digital image capturing device 831, a second digital image capturing device 832, a third digital image capturing device 833, a fourth digital image capturing device 834. The first digital image capturing device 831 may be centered at a position near the first end ipd.1 of the inter-pupillary distance width IPD, the fourth digital image capturing device 834 may be centered at a position near the second end ipd.2 of the inter-pupillary distance width IPD, and the remainder digital image capturing devices such as the second digital image capturing device 832 and the third digital image capturing device 833 may be positioned between or uniformly spaced apart from the first end ipd.1 and the second end ipd.2 of the inter-pupillary distance width IPD. In one embodiment, each digital image capturing device 831 through 834 or lens 818 may be surrounded by hypotenuse 814, surrounded by ring 816, and/or covered by lens cover 820, with lens 818 below lens cover 820.
It is further contemplated herein that the digital image capturing devices 831 through 834 may be used to capture a series of 2D images of the scene S.
Referring now to FIG. 8D, by way of example and not limitation, a cross-sectional view of an exemplary implementation of the image capture module 830 of FIG. 8C is illustrated. The image capturing module 830 may include digital image capturing devices 831 to 834, in which four image capturing devices are linearly connected in series within an intra-eye or inter-pupillary distance width IPD (distance between pupils of an average person). The digital image capturing devices 831 to 834 may include a first digital image capturing device 831, a second digital image capturing device 832, a third digital image capturing device 833, a fourth digital image capturing device 834. Each digital image capturing device 831 through 834 or lens 818 may be surrounded by hypotenuse 814, surrounded by ring 816, and/or covered by lens cover 820, with lens 818 below lens cover 820. It is contemplated herein that the digital image capturing devices 831 through 834 may include: an optical module, such as a lens 818, the lens 818 being configured to focus light from the scene S on a sensor module, such as: image capture sensor 822 the image capture sensor 822 is configured to generate an image signal for a captured image of scene S; and a data processing module 824, the data processing module 824 being configured to generate image data for capturing an image based on the generated image signal from the image capturing sensor 822.
It is contemplated herein that other sensor assemblies 822 may be utilized to generate image signals for captured images of scene S and other data processing modules 824 may be utilized to process or manipulate the image data.
It is contemplated herein that when the sensor 840 is not utilized to calculate different depths in the scene S (distances from the digital image capturing devices 831 through 834 to the foreground, background, and person P or object, such as the nearest point CP, key subject point KS, and the farthest point FP, as shown in fig. 4), then the user may be prompted to capture scene S images from the set distances of the digital image capturing devices 831 through 834 to the key subject point KS in the scene S, including but not limited to a 6 foot (6 ft) distance from the nearest point CP or key subject KS point of the scene S.
It is further contemplated herein that the digital image capturing devices 831 through 834 may be used to capture a series of 2D images of the scene S.
Referring now to FIG. 8E, by way of example and not limitation, an exploded view of an exemplary embodiment of an image capture module 830 is shown. The image capturing module 830 may include digital image capturing devices 831 to 833, in which a plurality or three image capturing devices are linearly connected in series within an intra-eye or inter-pupil distance width IPD (distance between pupils of an average person). The digital image capturing devices 831 to 833 may include a first digital image capturing device 831, a second digital image capturing device 832, and a third digital image capturing device 833. The first digital image capturing device 831 may be centered on a position near the first end ipd.1 of the inter-pupillary distance width IPD, the third digital image capturing device 833 may be centered on a position near the second end ipd.2 of the inter-pupillary distance width IPD, and the remaining image capturing devices such as the second digital image capturing device 832 may be centered on a center line CL between the first end ipd.1 and the second end ipd.2 of the inter-pupillary distance width IPDE. In one embodiment, each digital image capturing device 831 through 834 or lens 818 may be surrounded by hypotenuse 814, surrounded by ring 816, and/or covered by lens cover 820, with lens 818 below lens cover 820.
It is further contemplated herein that the digital image capturing devices 831 through 833 may be used to capture a series of 2D images of the scene S.
Referring now to FIG. 8F, by way of example and not limitation, a cross-sectional view of an exemplary embodiment of the image capture module 830 of FIG. 8E is illustrated. The image capturing module 830 may include digital image capturing devices 831 to 833, in which three image capturing devices are linearly connected in series within an intra-eye or inter-pupil distance width IPD (distance between pupils of an average person). The digital image capturing devices 831 to 833 may include a first digital image capturing device 831, a second digital image capturing device 832, and a third digital image capturing device 833. Each digital image capturing device 831 through 833 or lens 818 may be surrounded by hypotenuse 814, surrounded by ring 816, and/or covered by lens cover 820, with lens 818 below lens cover 820. It is contemplated herein that the digital image capturing devices 831 through 833 may include: an optical module, such as a lens 818, the lens 818 being configured to focus an image of the scene S on a sensor module, such as: an image capture sensor 822, the image capture sensor 822 configured to generate an image signal for a captured image of the scene S; and a data processing module 824, the data processing module 824 being configured to generate image data for capturing an image based on the generated image signal from the image capturing sensor 822.
Referring now to FIG. 8G, by way of example and not limitation, an exploded view of an exemplary embodiment of an image capture module 830 is shown. The image capturing module 830 may include a plurality of or two digital image capturing devices 831 to 832, in which the two image capturing devices are linearly connected in series within an intra-eye or inter-pupil distance width IPD (distance between pupils of an average person). The image capturing devices 831 to 832 may include a first image capturing device 831 and a second image capturing device 332. The first image capturing device 831 may be centered near a first end ipd.1 of the inter-pupillary distance width IPD and the second image capturing device 832 may be centered near a second end ipd.2 of the inter-pupillary distance width IPD. In one embodiment, each image capturing device 831-832 or lens 818 may be surrounded by hypotenuse 814, surrounded by ring 816, and/or covered by lens cover 820, with lens 818 below lens cover 320.
Referring now to FIG. 8H, by way of example and not limitation, a cross-sectional view of an exemplary embodiment of the image capture module 830 of FIG. 8G is shown. The image capturing module 830 may include digital or image capturing devices 831 to 832, wherein the two image capturing devices are linearly connected in series within an intra-eye or inter-pupil distance width IPD (distance between pupils of an average person). The image capturing devices 831 to 832 may include a first image capturing device 831 and a second image capturing device 332. Each image capturing device 831-832 or lens 818 may be surrounded by hypotenuse 814, surrounded by ring 816, and/or covered by lens cover 820, with lens 818 below lens cover 320. It is contemplated herein that the image capturing devices 831 through 832 may include: an optical module, such as a lens 818, the lens 818 being configured to focus an image of the scene S on a sensor module, such as: an image capture sensor 822, the image capture sensor 822 configured to generate an image signal for a captured image of the scene S; and a data processing module 824, the data processing module 824 being configured to generate image data for capturing an image based on the generated image signal from the image capturing sensor 822.
It is contemplated herein that other sensor components may be utilized to generate image signals for a captured image of scene S and other data processing modules 824 may be utilized to process or manipulate the image data.
It is contemplated herein that the image capturing module 830 and/or the digital image capturing devices 831 through 834 are used to obtain an offset 2D digital image view of the scene S. Further contemplated herein, image capture module 830 may include a plurality of image capture devices other than the numbers set forth herein, provided that the plurality of image capture devices are positioned within about an intra-ocular or inter-pupillary distance width IPD (interpupillary distance of an average person). In addition, it is further contemplated herein that image capture module 830 may include a plurality of image capture devices positioned within a linear distance approximately equal to inter-pupillary distance width IPD. Furthermore, it is further contemplated herein that image capture module 830 may include multiple image capture devices positioned vertically (computer system 10 or other smart device or portable smart device having short edge 811), multiple image capture devices positioned horizontally (computer system 10 or other smart device or portable smart device having long edge 812), or multiple image capture devices that are otherwise linearly spaced apart and within a linear distance approximately equal to inter-pupillary distance width IPD.
It is further contemplated herein that image capture module 830 and digital image capture devices 831 through 834 positioned linearly within the intraocular or interpupillary distance width IPD enable the rendering of accurate scene S in display 628 to produce a multi-dimensional digital image on display 628.
Referring now to fig. 9, by way of example and not limitation, a front view of a person's face is shown having left and right eyes LE, RE and each having a midpoint P1, P2 of the pupil to illustrate the inter-eye distance or intra-eye or inter-pupil distance IPD width—the distance between pupils of a typical person's vision system. Interpupillary distance (IPD) refers to the distance between the centers of pupils of the eyes measured in millimeters/inch. This measurement varies from person to person and also depends on whether they are looking at near or far objects. P1 may be represented by a first end ipd.1 of the inter-pupillary distance width IPD and PS may be represented by a second end ipd.2 of the inter-pupillary distance width IPD. The inter-pupillary distance IPD is preferably the distance between pupils of an average person and may have a distance between about two inches and a half-2.5 inches (6.35 cm), more preferably between about 40mm and 80mm, with the IPD of most adults being in the range of 50mm to 75mm, a wider range of 45mm to 80mm likely including (almost) all adults, and the minimum IPD of children (as low as 5 years) being about 40 mm.
It is contemplated herein that the left and right images may be generated as set forth in fig. 6.1-6.3 of us patent 9,992,473, us patent 10,033,990, and us patent 10,178,247 and electronically transferred to the left and right pixels 550L and 550R. In addition, the 2D image may be electronically transferred to the center pixel 550C.
Referring now to fig. 10, by way of example and not limitation, a representative illustration of a comfort circle (CoC) is shown to scale in fig. 4 and 3. For a defined plane, if a significant portion of the image is captured within a comfort circle (CoC), then the image captured on the lens plane will be comfortable and compatible with the human visual system of the user U viewing the final image displayed on display 628. Any object captured within the comfort circle (CoC), such as near plane N, key subject plane KSP, and far plane FP, by two image capturing devices such as image capturing devices 831 through 833 or image capturing devices 831 through 834 (interpupillary distance IPD), will be in focus for the viewer when rendered as a sequence of digital multi-dimensional images viewable on display 628. The backside object plane or far plane FP may be defined as the distance from the intersection of a 15 degree radial line to a vertical line in the field of view to a 30 degree line or R, which is the radius of the comfort circle (CoC). Further, a comfort circle (CoC) is defined as a circle formed by passing the diameter of the circle along the perpendicular to the key body KS plane (KSP), with the width being determined radially by 30 degrees of the center point on the lens plane-image capturing module 830-.
Image capturing devices such as digital image capturing devices 831 through 833 or digital image capturing devices 831 through 834 (interpupillary distance IPD) can be utilized to linear position or spacing within a 30 degree line just tangent to a comfort circle (CoC) to create motion parallax between multiple images when viewing a sequence of digital multi-dimensional images viewable on display 628, which would be comfortable and compatible with the human visual system of user U.
Referring now to fig. 10A, 10B, 10C and 11, by way of example and not limitation, right triangles resulting from fig. 10 are shown. All definitions are based on maintaining right triangles within the scene's relationship to image capture. Thus, knowing the key subject KS distance (convergence point), we can calculate the following parameters.
Fig. 6A is used to calculate the radius R of comfort (CoC).
R/KS=tan 30°
R=KS*tan 30°
Fig. 6B is used to calculate the optimal distance (inter-pupil distance IPD) between image capturing devices such as the image capturing devices 831 to 833 or the image capturing devices 831 to 834.
TR/KS=tan 15°
Tr=ks tan 15 °; and IPD is 2 x TR
Calculate the best far plane FP as FIG. 6C
Tan 15°=R/B
B=(KS*tan 30°)/tan 15°
Ratio of near plane NP to far plane fp= ((KS/(KS 8tan 30 °))) tan 15 °
To understand the meaning of TR, a 15 degree line hits/touches comfort (CoC) at a point on the linear image capture line of the lens plane. The images are arranged such that the key subject KS points are the same in all images captured by the image capturing devices such as the digital image capturing devices 831 to 833 or the digital image capturing devices 831 to 834.
The image capturing devices such as digital image capturing devices 831 to 833 or the users of digital image capturing devices 831 to 834 compose a scene S and in our case move digital image capturing device 830 such that the circle of confusion conveys scene S. Since the digital image capturing device 830 uses linearly spaced-apart multi-cameras, binocular parallax exists between multiple images or frames captured by digital image capturing devices 830 such as digital image capturing devices 831 through 833 or linear offsets of digital image capturing devices 831 through 834. This parallax can be changed by: changing the settings of the digital image capturing device 830 either brings the key subject KS back or away from the digital image capturing device to reduce parallax, or brings the key subject KS closer to the digital image capturing device to increase parallax. Our system is a fixed digital image capture device system and as a guideline for experimental development, the near plane NP should be no less than about 6 feet from the digital image capture device 830.
Referring now to FIG. 12, there is illustrated a flowchart step 1200 of a method performed by the computer system 10 and viewable on a display 628, the method having: capturing a plurality of 2D images of the scene S, generating frames 1101 to 1104, manipulating, reconfiguring, processing, displaying, storing a sequence of digital multidimensional images. Note that in fig. 12, some steps of designating a manual mode of operation may be performed by user U, where the user selects among the steps and provides input to computer system 10, whereas otherwise the operation of computer system 10 is based on the steps performed by application 624 in an automatic mode.
In block or step 1210, a computer system 10 is provided, the computer system 10 having a digital image capturing device 830, a display 628 and an application 624 as described above in fig. 1-11, such that a plurality of two-dimensional (2D) images having parallax due to the spacing of the digital image capturing devices 831-833, the digital image capturing devices 831-834, etc., within about an intra-eye or inter-pupil distance width IPD (distance between pupils of an average person) are captured and the computer system 10 displays a three-dimensional (3D) image sequence on the display 628. Further, digital images (DIFY or stereoscopic 3D) are sequentially displayed on the display 628, wherein images (n) among a plurality of 2D images of the scene S captured by the capturing devices 831 to 834 (n devices) are sequentially displayed on the display 628 as a digital multi-dimensional image sequence (DIFY or stereoscopic 3D).
In block or step 1215, the computer system 10 is configured, via the image capture application 624 (capture method), to capture a plurality of digital images of the scene S by capturing a plurality of 2D digital source images via an image capture module 830, the image capture module 830 having a plurality of image capture devices such as digital image capture devices 831-834, 830 (n devices) that are positioned linearly within an intra-ocular or inter-pupillary distance width IPD (distance between pupils of a human visual system within a comfortable circular relationship for optimizing digital multidimensional images of the human visual system), and the like. The computer system 10 integrates the I/O device 632 with the computer system 10, the I/O device 632 may include one or more sensors 840 in communication with the computer system 10 to measure the distance between the computer system 10 (image capturing devices, such as digital image capturing devices 831-834, 830 (n devices) and a selected depth (depth) in the scene S, such as a key subject KS, and to set the focus of the one or more digital image capturing devices 831-834, 830 (n devices).
The 3D stereoscopic, the user U may click or otherwise interact with the selection box 812 to select or identify key subjects KS in the source image pair, left image 1102, and right image 1103 of the scene S, as shown in fig. 16. Further, in block or step 1215, settings of the computer system 10, display 628, and application 206 (via the image capture application) are utilized to align or position icons, such as the cross-hair 814 of FIG. 16B, on the key subject KS of the scene S displayed on the display 628, for example by touching or dragging an image of the scene S, or pointing the computer system 10 in a different direction to align the cross-hair 814 of FIG. 16 on the key subject KS of the scene S. In block or step 1215, an image (n) focused on an image of the scene S or a selected depth of the scenes (depth) S is obtained or captured.
Alternatively, the computer system 10, through the image manipulation application 624 and the display 628, may be configured to operate in an automatic mode, wherein one or more sensors 840 may measure distances between the computer system 10 (image capturing devices, such as digital image capturing devices 831-834, 830 (n devices) and a selected depth (depth) in the scene S, such as the key subject KS).
It is recognized herein that user U may be instructed to: capturing, by the computer system 10, an image (n) of the scene S through the image capture application 624 and the display 628, for example, framing the scene S to include key subjects KS in the scene S, selecting salient foreground features of the scene S, and the furthest point FP in the scene S, may include identifying key subjects KS in the scene S, selecting closest points CP in the scene S, salient background features of the scene S, and so forth. Further, the key subjects KS in the scene S are positioned at a specific distance from the digital image capturing devices 831 to 834 (n devices). Further, the closest point CP in the scene S is located a specific distance from the digital image capturing devices 831 to 834 (n devices).
Referring now to FIG. 13, by way of example and not limitation, a touch screen display 628 is illustrated that enables user U to select a photography option of computer system 10. A first exemplary option may be DIFY capture, where user U may specify or select a digital image speed setting 1302, where user U may increase or decrease the playback speed or frames (images) per second of sequential display of digital images captured by capture devices 831 through 834 (n devices) on display 628. Further, the user U may designate or select the number of loops or repetitions 1304 of the digital image to set the number of loops of the image (n) 1000 of the plurality of 2D images of the scene S captured by the capturing devices 831 to 834 (n devices), 830 (n devices), wherein the image (n) 1000 of the plurality of 2D images of the scene S captured by the capturing devices 831 to 834 (n devices), 830 (n devices) is displayed on the display 628 in order, similar to fig. 11. Still further, the user U may designate or select a playback order 1306 of the digital image sequence or the palindromic sequence for playback to set a display order of the image (n) 1000 among the plurality of 2D images of the scene S captured by the capturing devices 831 to 834 (n devices), 830 (n devices). The timing sequence of the images shows that the proper binocular disparity is produced by the motion chase ratio effect. It is contemplated herein that the computer system 10 and the application 624 may utilize default or automatic settings herein.
Referring now to FIG. 16B, by way of example and not limitation, a touch screen display 628 is illustrated that enables user U to select a photography option (3D stereo) of computer system 10. In block or step 1215, settings of the computer system 10, the display 628, and the application 624 (via the image capture application) are utilized to align or position an icon, such as a reticle 1310, on the key subject KS of the scene S displayed on the display 628, for example by touching or dragging an image of the scene S, or to direct the computer system 10 in a different direction to align the reticle 1310 on the key subject KS of the scene S. In block or step 1215, an image (n) of scene S focused at a selected depth of the images or scenes (depths) of scene S is obtained or captured from image capturing devices 831 through 834 (n devices). The user U may click or otherwise interact with the selection box 1312 to select or identify a source image pair, a left image 1102L and a right image 1103R of a scene S selected from among the images (n) 1101, 1102, 1103, 1104 (frame set 1100) of the scene S of the image capturing devices 831 to 834 (n devices), or a key subject KS in any combination of two of the images (n) 1101, 1102, 1103, 1104 (frame set 1100). Moreover, the computer system 10, via the image manipulation application and display 624, may be configured to enable the user U to select or identify images of the scene S as left image 1102 and right image 1103 of the scene S. The user U may click on or otherwise interact 812 with the selection box to select or identify key subjects KS in the source image pair, left image 1102, and right image 1103 of the scene S, as shown in fig. 16B.
Alternatively, in block or step 1215, user U may utilize computer system 10, display 628, and application 624 to input multiple images, files, and datasets (Dataset) of scene S, such as through AirDrop, DROP OX, or other applications.
It is recognized herein that step 1215, computer system 10, through image capture application 624, image manipulation application 624, image display application 624, may be performed with different and separately located computer systems 10, such as one or more user systems 720, first smart device 722, second smart device 724, smart device(s), and application(s) 624. For example, using a camera system remote from the image manipulation system and remote from the image browsing system, step 1215 may perform approaching scene S through computer system 10 (first processor) and application 624 communicating between user systems 720, 722, 724 and application 624. Here, the camera system may be positioned or fixed to capture segments of different viewpoints of an event or entertainment, such as scene S. Next, the computer system 10 and the application 624 can capture and transmit a plurality of digital images of the scene S from the capture devices 831 through 834 (n devices) as a digital multi-dimensional image sequence (DIFY) of the scene S, a set of scene S images (n) with respect to the key subject KS point through the more user systems 720, 722, 724 via the communication link 740 and/or the network 750 or 5G.
The images captured at or near the inter-pupil distance IPD are matched to the human visual system, which simplifies mathematical operations, minimizes cross-talk between the two images, reduces blur and image movement, to produce a digital multi-dimensional image sequence (DIFY) that can be viewed on the display 628.
Further, in block or step 1215, settings of the computer system 10, the display 628, and the application 624 (via the image capture application) are utilized to align or position icons on the key subject KS of the scene S displayed on the display 628, such as the cross-hair 1310 of fig. 13 or 16B, for example, by touching or dragging a dataset of the scene S, or touching and dragging the key subject KS, or pointing the computer system 10 in different directions to align the cross-hair 1310 of fig. 13 or 16B on the key subject KS of the scene S. In block or step 1215, a Dataset (Dataset) of the scene S focused on a selected depth in an image or scene (depth) of the scene S is obtained or captured from a plurality of capture devices 830 (n devices).
Further, in block or step 1215, the I/O device 632 is integrated with the computer system 10, the I/O device 632 may include one or more sensors 852 in communication with the computer system 10 to measure the distance between the computer system 10/capture device 830 (n devices) and a selected depth in the scene S (depth), such as the key subject KS, and to set the focus of the arc or trajectory of the vehicle 400 and the capture device 830. It is contemplated herein that the computer system 10, display 628, and application 624 may operate in an automatic mode, wherein one or more sensors 840 may measure selected depths in the scene S (depth), such as the distance between key subjects KS. Alternatively, in manual mode, the user may determine the correct distance between the user U and a selected depth in the scene S (depth), such as the key subject KS. Or the computer system 10, the display 628 may utilize one or more sensors 852 to measure the distance between the capture device 830 (n devices) and a selected depth (depth) in the scene S, such as the key subject KS, and provide an indication or message (distance preference) on the screen to instruct the user U to move the capture device 830 (n devices) closer or farther from the key subject KS or the near plane NP to optimize the image, file, and Dataset (Dataset) of the capture device 830 (n devices) and the scene S.
In block or step 1220, the computer system 10, via the image manipulation application 624, is configured to receive, via the image acquisition application, a plurality of images of the scene S captured by the digital image capturing devices 831 through 834 (n devices), 830 (n devices). The image acquisition application converts each image into a digital source image, such as JPEG, GIF, TIF format. Ideally, each digital source image includes some visible objects, subjects, or points therein, such as foreground or nearest points associated with the near plane NP, the far plane FP, or furthest points associated with the far plane FP, and a key subject KS. The near plane NP and far plane FP points are the closest and farthest points, respectively, from the viewer (the plurality of capturing devices 831 and 832, 833 or 834, 830 (n devices)). Depth of field is the depth or distance (distance between the depicted foreground to background) formed within the object field. The principal axis is a line perpendicular to the scene passing through the key subject KS point, while the disparity is the displacement of the key subject KS point from the principal axis, see fig. 11. In digital synthesis, the shift is always kept as an integer number of pixels from the principal axis.
It is recognized herein that step 1220, computer system 10, through image capture application 624, image manipulation application 624, image display application 624, may be performed with different and separate computer systems 10, such as one or more user systems 720, 722, 724 and application programs 624. For example, using an image manipulation system that is remote from the image capture system and remote from the image viewing system, step 1220 may be performed remotely from scene S by computer system 10 (third processor) and an application that communicates between user systems 720, 222, 224 and application 624. Next, a set of multiple images (n) of the scene S may be received from the capturing devices 831 through 834 (n devices) relative to the key subject KS point through the communication link 740 and/or the network 750, or the 5G computer system 10 (third processor), or the application 624 through the more user systems 720, 722, 724, and the multiple digital multi-dimensional image sequences (DIFY or 3D stereoscopic) of the manipulated scene are transmitted to the computer system 10 (first processor) and the application 624 (step 1220A).
In block or step 1220A, the computer system 10 is configured, via an automatic key subject selection algorithm or key subject application 624, to identify a key subject KS in each source image, multiple images, of the scene S captured by the digital image capturing devices 831 through 834 (n devices). Further, the computer system 10 is configured to identify, via the key subject, the application 624, pixels, groups of pixels (finger point selections on the display 628) in one or more images (n) of the scene S from the digital image capturing devices 831 through 834 (n devices), respectively, as the key subject KS. Further, the computer system 10 is configured by the key subject application 624 to horizontally align a plurality of images of the source image, the scene S captured by the digital image capturing devices 831 to 834 (n devices), at a distance around the key subject KS; (horizontal image panning (HIT), as shown in fig. 11A and 11B, the key subject KS is within a comfortable circular relationship in order to optimize the digital multi-dimensional image/sequence 1010 of the human visual system.
Further, key subject points in a series of 2D images of the scene S are identified and each of the series of 2D images of the scene is aligned with a key subject KS point and all other points in the series of 2D images of the scene are shifted based on the spacing of the plurality of digital image capturing devices to generate a modified 2D image sequence or modified 2D image pair.
As shown in fig. 11A, 11B, and 4, in a plurality of images of a scene S each captured by the digital image capturing devices 831 to 834 (n devices), key subjects KS corresponding to the same key subject KS of the scene S can be identified. It is contemplated herein that the computer system 10, display 628, and application 624 may perform an algorithm or set of steps to automatically identify the subject KS in the plurality of images of the scene S captured by the digital image capturing devices 831 through 834 (n devices). Alternatively, in block or step 1220A, the computer system 10 (in manual mode—manual key subject selection algorithm), the display 628, and the settings of the application 624 are utilized to at least partially align or edit the user U with the pixels, groups of pixels (finger point selections), key subject KS points of at least two images (n) of the plurality of images of the scene S captured by the digital image capturing devices 831 through 834 (n devices).
The source image of the scene S, the plurality of images of the scene S captured by the digital image capturing devices 831 to 834 (n devices), are all obtained with the digital image capturing devices 831 to 834 (n devices) at the same image capturing distance and the same focal length. The computer system 10 creates a certain point, i.e., a key subject KS point, by the key subject application 624 by horizontally image shifting the source image, the plurality of images of the scene S captured by the digital image capturing devices 831 to 834 (n devices), at which the plurality of images of the scene S captured by the digital image capturing devices 831 to 834 (n devices) overlap, as shown in fig. 13. This shifting of the image does two things, first it sets the depth of the image. All points in front of the key subject KS point are closer to the viewer, and all points behind the key subject KS point are farther from the viewer.
Further, in the automatic mode, the computer system 10, through the image manipulation application, can identify the key subject KS based on the source image, the depth map of the plurality of images of the scene S captured by the digital image capturing devices 831 to 834 (n devices).
The computer system 10, through the image manipulation application, may use the source image, depth maps of multiple images of the scene S captured by the digital image capturing devices 831 through 834 (n devices), to identify the foreground, nearest and background, furthest points. Alternatively, in manual mode, the computer system 10, through the image manipulation application and display 628, may be configured to enable the user U to select or identify a key subject KS from among the source image of the scene S, the plurality of images of the scene S captured by the digital image capturing devices 831 through 834 (n devices), and the user U may click, move a cursor or box or other identification to select or identify the key subject KS from among the source image of the scene S, the plurality of images of the scene S captured by the digital image capturing devices 831 through 834 (n devices), as shown in fig. 13.
Horizontal Image Translation (HIT) sets key body plane KSP as the screen plane (first plane or near plane) that produces the scene. This step also sets the movement of objects, such as the movement of the shrubs B of the near plane NP (third plane or near plane) and the trees T of the far plane FP (second plane or far plane) relative to each other. The objects in front of the key body KS or key body plane KSP move in one direction (left to right or right to left), while the objects behind the key body KS or key body plane KSP move in the opposite direction to the objects in front. Objects behind the key body plane KSP will have less parallax in a given motion.
In the example of fig. 11, 11A and 11B, each layer 1100 includes the main image elements of the input file image of the scene S, such as images or frames 1101, 1102, 1103 and 1104 from digital image capturing devices 831 to 834 (n devices), respectively. The image acquisition application 624 performs a process to translate the images or frames 1101, 1102, 1103 and 1104, the images or frames 1101, 1102, 1103 and 1104 being overlapping and offset from the principal axis 1112 by the calculated disparity value, (horizontal image translation (HIT). Disparity line 1107 represents the linear displacement of key subject KS points 1109.1 to 1109.4 from the principal axis 1112. It is preferred that delta 1120 between disparity lines 1107 represents the linear amount of disparity 1120, such as front disparity 1120.2 and rear disparity 1120.1.
The disparities, minimum disparities, and maximum disparities are calculated from a function of pixel count, pixel density, and frame number, as well as the closest and farthest points and other parameters as set by U.S. patent 9,992,473, U.S. patent 10,033,990, and U.S. patent 10,178,247, which are all incorporated herein by reference.
In block or step 1220B, the computer system 10 is configured by the depth map application 624 to create a depth map of the source image, a plurality of images of the scene S captured by the digital image capturing devices 831 through 834 (n devices), and to make a grayscale image by an algorithm. The depth map is an image or image channel that contains information about the distance of objects, surfaces or points in the scene S from a point of view, such as digital image capturing devices 831 through 834 (n devices). This provides more information, for example, because volume, texture, and illumination are more fully defined. Once depth map 1220B is generated, the disparity may be tightly controlled. To this end, the computer system 10 may limit the number of output frames to four without turning to the depth map. If we use four from the depth map or two from the depth map we are not constrained by the middle camera position. Note that the external digital image capturing devices 831 and 834 are locked into the interpupillary distance (IPD) of the viewer or user U viewing display 628.
In addition, the computer system 10, via the key subject, application 624, may identify the key subject KS based on the depth map of the source image. Similarly, computer system 10, via depth map application 624, may use the depth map of the source image to identify a near plane NP, a far plane FP, a foreground, a nearest point and a background, a farthest point, the near plane NP may be a plane through the focus closest to the lens plane (brush B in the foreground), and the far plane FP is a plane through the farthest focus (tree T in the background).
The computer system 10, via the depth map application 624, may define two or more planes for each of a series of 2D images of the scene, and one or more planes may have different depth estimates. Computer system 10, through depth map application 624, may identify a first proximal plane, such as key subject plane KSP, and a second distal plane, such as near plane NP or far plane FP, within a series of 2D images of a scene.
In block or step 1220C, the computer system 10 is configured via the plug-in intermediary application 624 to superimpose a model or grid of the image (n) 1101, 1102, 1103, 1104 (frame set 1100) of the scene S with RGB high resolution colors (DIFY or 3D stereoscopic) on the 2D RGB high resolution digital camera.
In block or step 1220D, the computer system 10, via the dataset manipulation application 624, may be used to generate a model or grid of the scene S from the image (S) 1101, 1102, 1103, 1104 (the set of frames 1100) of the scene S.
In block or step 1225, the computer system 10 is configured by the frame creation program 624 to create or generate frames, record images of the (n) 1101, 1102, 1103, 1104 (frame set 1100) of the scene S from the virtual camera movement, rotation, or arc position, such as a separation or movement such as-5, -4, -3, -2, -1, 0, 1, 2, 3, 4, 5, such as 0.5 to 1 degree between frames; generating disparities for 3D color grid data sets 1101, 1102, 1103, 1104 (frame set 1100) of the DIFY representative scene S; for left and right 3D stereoscopic; such as 1102, 1103 images of the 3D color grid dataset and 1101, 1102, 1103, 1104 of the (DIFY). The computer system 10, via the key body, application 624, may establish an increment of offset, such as one (1) degree of total offset between views (typically offset by 10-70 pixels on the display 628). This simply means that a 360 degree full sensor (capture device) rotation around the key subject KS will have a 360 degree view, so we only use/need view 1 and view 2, left and right for 3D stereo; 1102. 1103 image dataset. Assuming that the rotational disparity spirals around the key subject (zero disparity point), this provides a 1 degree separation/disparity for each view. This will likely establish a minimum disparity/parallax that can be adjusted upward as the sensor (image capture module 830) moves farther away from the key subject KS.
In block or step 1225A, the computer system 10 is configured through the frame creation program 624 to input or upload source images captured from outside the computer system 10.
In block or step 1230, the plurality of images of the scene S captured by the digital image capturing devices 831 through 834 (n devices) may be configured to align or convert each source image horizontally and vertically by the horizontal and vertical frame DIF panning program 624 using the computer system 10, requiring a Dimensional Image Format (DIF) conversion. The DIF transform is a geometric shift that does not change the information obtained at each point in the source image, the plurality of images of the scene S captured by the digital image capturing devices 831 to 834 (n devices), but can be regarded as a shift in cartesian space of all other points in the source image, the plurality of images of the scene S captured by the digital image capturing devices 831 to 834 (n devices) (as illustrated in fig. 11). As a plenoptic function, the DIF transformation is represented by the following equation:
wherein Δu, v=Δθ, Φ
In the case of a digital image source, the geometric shift corresponds to the geometric shift of the pixels containing plenoptic information, and then the DIF transform becomes:
(Pixel) x,y =(Pixel) x,yx,y
in addition, computer system 10 may also geometrically shift the background and or foreground using the DIF transform by horizontal and vertical frame DIF translation application 624. The background and foreground may be geometrically shifted according to the depth of the key subject KS identified by the depth map 1220B of the plurality of images of the scene S captured by the digital image capturing devices 831 through 834 (n devices), each with respect to the source image. Controlling the geometric shift of the background and foreground relative to the key subject KS controls the motion parallax of the key subject KS. As noted, the apparent relative movement of the key subject KS to the background or foreground provides an observer with hints about their relative distance. In this way, motion parallax is controlled to focus objects at different depths in the displayed scene to match the parallax requirements of the vergence and stereoscopic retina, thereby better simulating natural viewing conditions. The cues for eyeball accommodation and convergence are consistent by adjusting the focus of the key subject KS in the scene to match its stereoscopic retinal disparity (the intra-ocular or inter-pupillary distance width IPD (distance between pupils of the human visual system)).
Referring again to fig. 4, viewing the DIFY, multi-dimensional image sequence 1010 on display 628 requires two different eye movements of user U. First, the eye will track the nearest item, point or object (near plane NP) in the multi-dimensional image sequence 1010 on display 628, which will linearly translate back and forth to the stationary key subject plane KSP as images or frames 1101, 1102, 1103 and 1104 are superimposed and offset from the principal axis 1112 by the calculated disparity value (horizontal image translation (HIT)). This tracking occurs through eye movement following motion. Second, the eye perceives depth due to smooth motion changes of any point or object relative to the key subject plane KSP, and more specifically, relative to the key subject KS point. Thus, DIFY consists of one mechanical step and two eye functions.
One mechanical step is to shift the frames so that the key subject KS points overlap over all frames. Since images or frames 1101, 1102, 1103 and 1104 may be overlapping and offset from the principal axis 1112 by the calculated disparity value (horizontal image translation (HIT)), the translation is linear back and forth to the stationary key body plane KSP. The eye follows the motion of the near-plane NP object, exhibiting the greatest motion (eye rotation) relative to the key subject KS. The difference in frame position along the key body plane KSP (smooth eye movement) introduces binocular disparity. Comparison of any two points other than the key subject KS also produces depth (binocular parallax). The point behind key body plane KSP moves in the opposite direction from the point in front of key body KS. Comparison of two points in front of or behind or across the key body KS plane shows depth.
In block or step 1235A, the computer system 10 is configured, via the palindromic application 626, to create, generate, or generate a multi-dimensional sequence of digital images 1010 with seamless palindromic loops, to sequentially align (sequentially align) each of the images (n) of the scene S from the digital image capturing devices 831 through 834 (n devices), such as sequentially display the following loops: a first digital image, image or frame 1101 (1) from a first digital image capturing device 831, a second digital image, image or frame 1102 (2) from a second digital image capturing device 832, a third digital image, image or frame 1103 (3) from a third digital image capturing device 833, and a fourth digital image, image or frame 1104 (4) from a fourth digital image capturing device 834. In addition, another sequence is the following loop: a first digital image, image or frame 1101 (1) from a first digital image capturing device 831, a second digital image, image or frame 1102 (2) from a second digital image capturing device 832, a third digital image, image or frame 1103 (3) from a third digital image capturing device 833, a fourth digital image, image or frame 1104 (4) from a fourth digital image capturing device 834, a third digital image, image or frame 1102 (3) from a third digital image capturing device 833, a second digital image, image or frame 1102 (2) from a second digital image capturing device 832, a first digital image, image or frame 1101 (1) -1, 2, 3, 4, 3, 2, 1 (aligned in sequence) from a first digital image capturing device 831. The preferred sequence is that the source image, the plurality of images of scene S captured by the digital image capturing devices 831 through 834 (n devices), follow the same order or sequence in which the images were captured, and add an inverted or reverse order to create a seamless palindromic loop.
It is contemplated herein that other sequences may be configured herein, including but not limited to 1, 2, 3, 4, 3, 2, 1 (aligned in sequence), etc.
It is contemplated herein that the first proximal plane is horizontally and vertically aligned, e.g., the key subject plane KSP of each image of the images (n) of the scene S from the digital image capturing devices 831 to 834 (n devices), and the second distal plane of each subsequent image frame in the sequence, e.g., the foreground plane, the near plane NP, or the background plane, the far plane FP, is shifted based on the depth estimate of the second distal plane to produce a second modified 2D image sequence or a second modified 2D image pair.
In block or step 1235B, the computer system 10 may be configured to inter-phase (inter) the pixel columns of each frame set 1100, specifically the left image 1102 and the right image 1103, by the inter-correspondence 626 to generate a multi-dimensional digital image that is point aligned with the key subject KS and within the calculated parallax range. As shown in fig. 16A, the inter-correspondence 626 may be configured to obtain segments, bands, rows, or columns of pixels from the left image 1102 and the right image 1103, such as a source image pair of the scene S, the left image 1102, and the right image 1103, and alternately layering them between the columns 1602A of the left image 1102-LE and the columns 1602A of the right image 1103-RE, reconfigure or arrange them side-by-side in a staggered series, such as a repeating series 160A of two columns wide, and repeat the configuration for all layers of the source image pair of the topography T of the scene S, the left image 1102, and the right image 1103, wherein the columns 1602A are sized to be one pixel 1550 wide.
It is contemplated herein that the source image, multiple images of scene S captured by capture device 830, matches the size and configuration of display 628, is aligned with the critical subject KS point and is within the calculated parallax range.
Now given the multi-dimensional image sequence 1010, we turn to view the viewing side of the device.
It is contemplated herein that the source image, multiple images of scene S captured by digital image capturing devices 831 through 834 (n devices), match the size and configuration of display 628, align with key subject KS points and are within the calculated parallax range.
In block or step 1240, computer system 10, via image editing application 624, is configured to crop, scale, align, enhance, or perform editing or editing of multi-dimensional digital image sequence 1010 on each image (n) of scene S from capturing devices 831 through 834 (n devices).
In addition, the computer system 10 and editing application 624 may enable the user U to perform frame enhancement, layer enrichment, animation, feathering (smoothing), (photo or Acorn photos or image tools), smoothing or filling in images (n), or other software techniques for creating 3D effects on the display 628. It is contemplated herein that the computer system 10 (auto mode), display 628, and application 624 may perform an algorithm or set of steps to automatically or cause automatic execution of the alignment or edit alignment, cropping, scaling, alignment, enhancement of the pixels, or editing of the key subject KS points or the multi-dimensional digital image sequence 1010 on the plurality of images of the scene S captured by the digital image capturing devices 831 through 834 (n devices).
Alternatively, in block or step 1240, the settings of computer system 10 (in manual mode), display 628, and application 624 are utilized to at least partially align or edit user U to the pixels of key subject KS points, the alignment of groups of pixels, cropping, scaling, aligning, enhancing, or performing editing or editing of the multi-dimensional digital image sequence 1010 on the plurality of images of scene S captured by digital image capturing devices 831 through 834 (n devices).
In addition, the user U can set or select the speed (viewing time) and the number of always viewing cycles for each frame through the display 628 and the editing application 624, as shown in fig. 13. A time interval may be allocated to each frame in the multi-dimensional digital image sequence 1010. In addition, the time interval between frames may be adjusted at step 1240 to provide a smooth motion and optimal 3D view of the multi-dimensional digital image sequence 1010.
It is contemplated herein that the computer system 10, the display 628, and the application 624 may perform an algorithm or a set of steps to automatically or manually edit or apply effects to the set of frames 1100. In addition, the computer system 10 and editing application 206 may include editing, such as frame enhancement, layer enrichment, eclosion, (photo or Acorn photo or image tools) to smooth or fill in the image (S) together, as well as other software techniques for producing 3D effects to display a 3D multidimensional image of the topography T of the scene S on the display 628.
In block or step 1250, the computer system 10 is configured, via the image display application 624, to cause the image (n) of the scene S to display the multi-dimensional digital image sequence 1010 of the scene S on the display 628 via a sequential palindromic loop for the different sized displays 628. Likewise, the multi-dimensional digital image sequence 1010 of scene S, the resulting 3D image sequence, may be output as a DIF sequence to the display 628. It is contemplated herein that computer system 10, display 628, and application 624 may be responsive, i.e., computer system 10 may execute instructions to size each image (n) of scene S to fit the size of a given display 628.
In addition, the user U may choose to return to block or step 1220 to select a new key subject KS among the plurality of images, each source image of the scene S captured by the digital image capturing devices 831 through 834 (n devices), and proceed through steps 1220 through 1250 to view the multi-dimensional digital image sequence 1010 of the scene S of the new key subject KS on the display 628 by creating a new or second sequential loop.
Now given the multi-dimensional image sequence 1010, we turn to view the viewing side of the device. Further, in block or step 735, the computer system 10, via the output application 730 (206), may be configured to display the multi-dimensional image 1010 on the display 628 for one or more user systems 220, 222, 224 via the communication link 240 and/or the network 250, or the 5G computer system 10 and the application program 206.
For 3D stereoscopic, referring now to fig. 15A, a cross-sectional view of an exemplary stack of components of a display 628 is shown by way of example and not limitation. The display 628 may include an array or plurality of pixels of emitted light, such as a stack of LCD panel components 1520 having electrodes (such as front and rear electrodes), polarizers (such as horizontal and vertical polarizers), diffusers (such as gray diffusers, white diffusers), and backlights that emit red R light, green G light, and blue B light. In addition, the display 628 may include other standard LCD user U interactive components, such as a top glass cover 1510, with a capacitive touch screen glass 1512 positioned between the top glass cover 1510 and the stack 1520 of LCD panel components. It is contemplated herein that other forms of display 628, such as LED, ELED, PDP, QLED and other types of display technologies, may be included herein in addition to an LCD. Further, the display 628 may include a lens array, such as a lenticular lens 1514 preferably positioned between the capacitive touch screen glass 1512 and the LCD panel component stack 1520, and configured to bend or refract light in the following manner: a pair of interlaced left and right stereoscopic image pairs can be displayed as a 3D or multi-dimensional digital image 1010 on the display 628 to thereby display a multi-dimensional digital image of the scene S on the display 628. Transparent adhesive 1530 can be used to bond components in a stack, whether used as a horizontal adhesive or a vertical adhesive to hold multiple components in a stack. For example, to generate a 3D view or to generate a multi-dimensional digital image on the display 628, it is necessary to divide 1920x1200 pixel images by a plurality of pixels into two halves, i.e., 960x1200, and any half of the plurality of pixels can be used for the left image and the right image.
It is contemplated herein that the lens array may include other techniques for bending or refracting light, such as barrier screens (black lines), biconvex, parabolic, overlay, waveguide, black lines, etc., that are capable of separating into left and right images.
It is further contemplated herein that the lenticular lenses 514 may be oriented in vertical columns to produce a multi-dimensional digital image on the display 628 while the display 628 remains in a landscape view. However, the 3D effect is not apparent when the display 628 remains in a portrait view, enabling 2D and 3D viewing using the same display 628.
It is further contemplated herein that smoothing or other image noise reduction techniques and foreground subject focusing may be used to soften and enhance the 3D view or multi-dimensional digital image on the display 628.
Referring now to fig. 15B, by way of example and not limitation, a representative segment or section of one embodiment of an exemplary refractive element, such as a lenticular 1514 of a display 628, is shown. Each subelement of the lenticular lens 1514 is an arcuate or curved or arched section or segment 1540 (shaped as an arc) of the lenticular lens 1514, and each subelement of the lenticular lens 514 can be configured with a repeating series of trapezoidal lens segments or a plurality of subelements or refractive elements. For example, each arcuate or curved or arched segment 1540 may be configured with a lens peak 1541 of the lenticular lens 1540 and sized to be one pixel 1550 (emitting red R light, green G light, and blue B light) wide, such as a center pixel 1550C with an assignment to the lens peak 1541. It is contemplated herein that the center pixel 1550C light passes through the lenticular lens 1540 as center light 1560C to provide 2D viewing of the image on the display 628 to the left eye LE and right eye RE of the viewing distance VD from the trapezoid segment or section 1540 of the pixel 1550 or lenticular lens 1514. Further, each arcuate or curved segment 1540 may be configured to have an angled segment, such as a lens angle A1 of a lens refractive element, such as lens sub-element 1542 (multiple sub-elements) of lenticular lens 1540, and sized to be one pixel wide, such as left pixel 1550L and right pixel 1550R with left lens sub-element 1542L having angle A1 and right lens sub-element 1542R having angle A1 assigned to the left lens, such as tilt angle and dip angle, respectively, of light refracted through centerline CL. It is contemplated herein that pixel 1550L/R light passes through lenticular 1540 and bends or refracts to provide left and right images to enable 3D viewing of the image on display 628; light passing through left pixel 1550L passes through left lens angle 1542L and bends or refracts, such as light entering left lens angle 1542L bends or refracts to pass through center line CL to the right R side, left image light 1560L is directed toward left eye LE, and right pixel 1550R light passes through right lens angle 1542R and bends or refracts, such as light entering right lens angle 1542R bends or refracts to pass through center line CL to the left side L, right image light 1560R is directed toward right eye RE, to produce a multi-dimensional digital image on display 628.
It is contemplated herein that the left and right images may be generated as set forth in fig. 6.1-6.3 of us patent 9,992,473, us patent 10,033,990, and us patent 10,178,247 and electronically transferred to the left and right pixels 550L and 550R. Further, the 2D image may be electronically transferred to the center pixel 550C.
In this figure, each lens peak 1541 has corresponding left and right angled lenses 1542, such as left and right angled lenses 1542L and 1542R on either side of the lens peak 1541 and each assigned one pixel, respectively, to which a center pixel 1550C, a left pixel 1550L, and a right pixel 1550R are assigned.
In this figure, viewing angle A1 is a function of viewing distance VD, size S of display 628, where a1=2 arctan (S/2 VD)
In one embodiment, each pixel may be configured by a set of subpixels. For example, to produce a multi-dimensional digital image on the display 628, each pixel may be configured as one or two 3x3 subpixels of the LCD panel assembly stack 1520, with the LCD panel assembly stack 520 emitting one or two red R lights, one or two green G lights, and one or two blue B lights through a segment or section of the lenticular 1540 to produce the multi-dimensional digital image on the display 628. The red R light, green G light, and blue B light may be configured as a vertical stack of three horizontal subpixels.
It is recognized herein that the trapezoidal lens 1540 bends or refracts light uniformly through its center C, left L side, and right R side, such as left angled lens 1542L and right angled lens 1542R and lens peak 1541.
Referring now to fig. 15C, a prototype section or segment of one embodiment of an exemplary lenticular 1514 of a display 628 is shown by way of example and not limitation. Each segment or multiple subelements or refractive elements that are trapezoidal segments or sections 1540 of the lenticular lens 1514 can be configured with a repeating series of trapezoidal lens segments. For example, each trapezoid segment 1540 may be configured with a lens peak 1541 of the lenticular lens 1540 and sized to be one or two pixels 1550 wide and a flat or straight lens such as lens valley 1543 and sized to be one or two pixels 1550 wide (emitting red R light, green G light, and blue B light). For example, the lens valleys 1543 may be assigned the center pixel 1550C. It is contemplated herein that the center pixel 1550C light passes through the lenticular lens 1540 as center light 1560C to provide 2D viewing of the image on the display 628 to the left eye LE and right eye RE of the viewing distance VD from the trapezoid segment or section 1540 of the pixel 1550 or lenticular lens 1514. Further, each trapezoid segment 1540 may be configured to have an angled portion, e.g., lens angle 1542 of lenticular lens 1540 and sized to be one or two pixels wide, such as left pixel 1550L and right pixel 1550R assigned to left lens angle 1542L and right lens angle 1542R, respectively. It is contemplated herein that pixel 1550L/R light passes through lenticular 1540 and curves to provide left and right images to enable 3D viewing of the image on display 628; light passing through left pixel 1550L passes through left lens angle 1542L and bends or refracts, such as light entering left lens angle 1542L bends or refracts to pass through center line CL to the right R-side, left image light 1560L toward left eye LE; and right pixel 1550R light passes through right lens angle 1542R and bends or refracts, such as light entering right lens angle 1542R bends or refracts to pass through center line CL to left side L, right image light 1560R toward right eye RE;
It is contemplated herein that the angle A1 of the lens angle 1542 is a function of the size of the pixel 1550, the stack of components of the display 628, the refractive characteristics of the lenticular lens 514, and the distance of the left and right eyes LE, RE from the pixel 1550—the viewing distance VD.
In this fig. 15C, the viewing angle A1 is a function of the viewing distance VD, the size S of the display 628, where a1=2 arctan (S/2 VD).
Referring now to fig. 15D, a representative segment or section of one embodiment of an exemplary lenticular 1514 of a display 628 is shown by way of example and not limitation. Each segment or multiple subelements or refractive elements that are parabolic or dome-shaped segments or sections 1540A (parabolic or dome lenses, domes) of the lenticular lens 1514 may be configured with a repeating series of dome-shaped, curved, semicircular lens segments. For example, each dome segment 1540A may be configured with a lens peak 1541 of the lenticular lens 540 and sized to be one or two pixels 1550 (emitting red R light, green G light, and blue B light) wide, such as a center pixel 1550C with an assignment to the lens peak 1541. It is contemplated herein that the center pixel 1550C light passes through the lenticular lens 540 as center light 560C to provide 2D viewing of the image on the display 628 to the left eye LE and right eye RE of the viewing distance VD from the trapezoidal section or segment 1540 of the pixel 1550 or lenticular lens 1514. Further, each trapezoid segment 1540 may be configured to have an angled portion, e.g., lens angle 1542 of a lenticular lens 1540 and sized to be one pixel wide, such as left pixel 1550L and right pixel 1550R assigned to left lens angle 1542L and right lens angle 1542R, respectively. It is contemplated herein that pixel 1550L/R light passes through lenticular 1540 and curves to provide left and right images to enable 3D viewing of the image on display 628; light passing through left pixel 1550L passes through left lens angle 1542L and bends or refracts, such as light entering left lens angle 1542L bends or refracts to pass through center line CL to the right R side, left image light 1560L is directed toward left eye LE, and right pixel 1550R light passes through right lens angle 1542R and bends or refracts, such as light entering right lens angle 1542R bends or refracts to pass through center line CL to the left side L, right image light 1560R is directed toward right eye RE to produce a multi-dimensional digital image on display 628.
It is recognized herein that dome lens 1540B bends or refracts light almost uniformly through its center C, left L side and right R side.
It is recognized herein that representative segments or sections of one embodiment of the exemplary lenticular lens 1514 may be configured in a variety of other shapes and sizes.
Furthermore, to achieve both highest quality two-dimensional (2D) image viewing and multi-dimensional digital image viewing on the same display 628, the digital form of alternating black lines or parallax barriers (alternating) may be used during multi-dimensional digital image viewing on the display 628 without adding lenticular lenses 1514 to the stack of displays 628, and then the digital form of alternating black lines or parallax barriers (alternating) may be disabled during two-dimensional (2D) image viewing on the display 628.
Parallax barriers are devices placed in front of an image source, such as a liquid crystal display, to allow it to display stereoscopic or multiview (multiview) images without the viewer wearing 3D glasses. Placed in front of a conventional LCD, it consists of a narrow opaque layer with a series of precise spacings, allowing each eye to see a different set of pixels, creating a sense of depth through parallax. A digital parallax barrier is a series of alternating black lines in front of an image source such as a liquid crystal display (pixel) to allow it to display stereoscopic or multi-vision images. Furthermore, the face tracking software function may be used to adjust the relative positions of the pixels and barrier slits according to the position of the user's eyes, allowing the user to experience 3D from a wide range of positions. KeehoonHong, soon-gi Park, jiroo Hong, byoungho Lee, book Design and Implementation of Autostereoscopic Displays (design and implementation of autostereoscopic displays) is incorporated herein by reference.
It is contemplated herein that parallax and key subject KS reference point calculations may be arrayed for distances between virtual camera locations, inter-phase spacing, distance of display 628 from user U, lenticular 1514 configuration (lens angles A1, 1542, lens per millimeter and millimeter depth), lens angle 1542 as a function of display 628 component stack, refractive characteristics of lenticular 1514, and distance of left eye LE and right eye RE from pixel 1550, viewing distance VD, distance location between virtual cameras (pupil spacing IPD), etc. to produce digital multi-dimensional images related to viewing devices or other viewing functions, such as barrier screens (black lines), lenticular lenses, parabolas, overlays, waveguides, black lines, etc. with integrated LCD layers in LEDs or OLED, LCD, OLED and combinations thereof or other viewing devices.
Incorporated herein by reference is the Jason Geng's paper entitled Three-Dimensional Display Technology (Three-dimensional display technology), other display technologies, etc., pages 1-80, which can be used to generate the herein display 628 incorporated by reference.
It is contemplated herein that the number of lenses per millimeter or inch of lenticular lenses 514 is determined by the number of pixels per inch of display 628.
It is contemplated herein that other angles A1 are contemplated herein; the distance (about 0.5 mm) of the pixels 1550C, 1550L, 1550R from the lens 1540; and the viewing distance of user U from the user's eyes (about fifteen (15) inches) from the smart device display 628), and the average human inter-pupillary distance between the human eyes (about 2.5 inches) can be decomposed or calculated to produce a digital multi-dimensional image. The control rules of angle and spacing ensure that the viewing image on display 628 is within the comfort zone of the viewing device to produce a digital multi-dimensional image, see figures 5, 6, 11 below.
It is recognized herein that the angle A1 of the lens 1541 may be calculated and set based on the viewing distance VD between the user U-eyes, left eye LE and right eye RE, and pixels 550, such as pixels 1550C, 1550L, 1550R, which is a comfortable distance to maintain the display 628 from the user U-eyes, such as ten (10) inches to the length of the arm/wrist, or more preferably between about fifteen (15) inches to twenty-four (24) inches, and most preferably about fifteen (15) inches.
In use, the user U moves the display 628 toward and away from the user ' S eyes until a digital multi-dimensional image appears in front of the user, which movement takes into account the user ' S U actual inter-pupillary distance IPD separation and matches the user ' S vision system (near-side difference and far-side difference), the width position of the left-right interleaved images as a function of distance from the virtual camera position (inter-pupillary distance IPD), the depth of the key subject KS in each of the digital images (n) of the scene S (key subject KS algorithm), the horizontal image panning algorithm for the key subject KS for both images (left and right images), the inter-dimensional algorithm for the key subject KS for both images (left and right images), the angle A1, the distance of the pixel 1550 from the lens 1540 (pixel-lens distance (PLD) of about 0.5 mm), and the refractive characteristics of the lens array such as trapezoidal lens 1540, all of which are considered to produce a digital image for the user U to view the display 628. The first known elements are the number of pixels 1550 and the number of images, the two images, the distance between the virtual camera positions or (interpupillary distance IPD). The image captured at or near the inter-pupil distance IPD matches the human visual system, which simplifies mathematical operations, minimizing cross-talk, blurring, image movement between the two images to produce a digital multi-dimensional image that can be viewed on the display 628.
It is further contemplated herein that the trapezoidal lenses 1540 may be formed of polystyrene, polycarbonate, or other transparent material or similar materials, as these materials provide a variety of forms and shapes, may be manufactured in different shapes and sizes, and provide reduced weight strength; however, other suitable materials, etc., may be used so long as such materials are transparent and are machinable or formable to meet the purposes described herein to produce left and right stereoscopic images and a specified refractive index. It is further contemplated herein that the trapezoidal shaped lens 1541 may be configured with 4.5 lenticules per millimeter and a depth of approximately 0.33 mm.
DIFY, in block or step 1250, computer system 10 is configured, via image display application 624, to cause frame set 1100 of topography T to display multi-dimensional digital image sequence 1010 on display 628 via sequential palindromic cycles for different sized displays 628. Likewise, the multi-dimensional digital image sequence 1010 of scene S, the resulting 3D image sequence, may be output to the display 628 as a DIF sequence or a. MPO file. It is contemplated herein that computer system 10, display 628, and application 624 may be responsive, i.e., computer system 10 may execute instructions to size each image (n) of scene S to fit the size of a given display 628.
In block or step 1250, the multi-dimensional image sequence 1010 on the display 628 utilizes the positional differences of the object of each of the images (n) of the scene S from the capturing devices 831 through 834 (n devices) relative to the key subject plane KSP, which differences introduce disparities between the images in the sequence to display the multi-dimensional image sequence 1010 on the display 628 to enable the user U to view the multi-dimensional image sequence 1010 of the display 628 in block or step 1250.
Further, in block or step 1250, the computer system 10, via the output application 624, may be configured to display the multi-dimensional image sequence 1010 on the display 628 via the communication link 740 and/or the network 750, or the 5G computer system 10 and the application 624 for one or more user systems 720, 722, 724.
In block or step 1250, computer system 10, via output application 624, may be configured to display multi-dimensional image 1010 on display 628. The multi-dimensional image 1010 may be displayed by left and right pixels 1102L/1103R, with light passing through lenticular 1540 and bending or refracting to provide 3D viewing of the multi-dimensional image 1010 on the display 628 to the left eye LE and right eye RE, a viewing distance VD from pixel 1550.
In block or step 1250, each image (n) (L & R segments) of scene S of frame set 1100 from topography T of scene S is configured with computer system 10, display 628 and application 624 settings while the key subject is aligned between the images for binocular disparity for displaying/viewing/saving multi-dimensional digital master image 1010 on display 208, wherein the positional differences of each image (n) of scene S from the virtual camera introduce (left and right) binocular disparity relative to the key subject KS plane to display multi-dimensional digital image 1010 on display 208 to enable user U to view the multi-dimensional digital image on display 208 in block or step 1250.
In addition, the user U may choose to return to block or step 1220 to select a new key subject KS in the set of frames 1100 of each source image, topography T of the scene S, and proceed through steps 1220 through 1250 to view the multi-dimensional digital image sequence 1010 of the scene S of the new key subject KS on the display 628 by creating a new or second sequential loop.
The display 628 may include a display device (e.g., a viewing screen implemented on a smart phone, PDA, monitor, television, tablet, or other viewing device capable of projecting information in pixel format) or a printer (e.g., a consumer printer, kiosk, dedicated printer, or other hard copy device) to print a multi-dimensional digital master image on, for example, a raster or other physical viewing material.
It is recognized herein that steps 1220 through 1240 may be performed by computer system 10 through image manipulation application 626 using different and separately located computer systems 10, such as one or more user systems 720, 722, 724 and application programs 626, to perform the steps herein. For example, using an image processing system remote from the image capture system and remote from the image browsing system, steps 1220 through 1240 may be performed remotely from scene S by computer system 10 or server 760 and application 624, and communication may be performed between user systems 720, 722, 724 and application 626 through communications link 740 and/or network 750, or through a wireless network such as 5G computer system 10 and application 626 through more user systems 720, 722, 724. Here, the computer system 1, via the image manipulation application 624, can manipulate 24 settings to configure each image (n) (L & R segments) of the scene S from the scene S of the virtual camera to generate a multi-dimensional digital image sequence 1010 aligned with key subject KS points and transmitted to one or more user systems 720, 722, 724 via a communication link 740 and/or network 750 or via a wireless network such as the 5G computer system 10 or server 760 and application 624 to display the multi-dimensional digital image/sequence 1010.
Further, it is herein recognized that steps 1220 through 1240 may be performed by computer system 10 through image manipulation application 624 utilizing a different and separately located computer system 10 positioned on a vehicle. For example, using an image processing system remote from the image capture system, steps 1220 through 1240 are performed by computer system 10 and application 624, computer system 10 may manipulate 24 settings to configure each image (n) (L & R segment) image of scene S from scene S of capture device 830 to generate a multi-dimensional digital image/sequence 1010 aligned with key subject KS points. Here, the computer system 10, through the image manipulation application 626, may utilize the multi-dimensional image/sequence 1010 to navigate the vehicle V through the terrain T of the scene S. Alternatively, computer system 10, through image manipulation application 626, may cause user U remote from vehicle V to use multi-dimensional image/sequence 1010 to navigate vehicle V through terrain T of scene S.
It is contemplated herein that computer system 10, via output application 624, may be configured to cause multi-dimensional image sequence 1010 to be displayed on display 628 such that a plurality of users U view multi-dimensional image sequence 1010 on display 628 in block or step 1250, either live or as a replay/rebroadcast.
It is recognized herein that step 1250 may be performed by computer system 10 through output application 624 utilizing a different and separately located computer system 10, such as one or more user systems 720, 722, 724 and application programs 624 performing the steps herein. For example, communications between user systems 720, 722, 724 and application 626 are made through computer system 10 and application 624 using output or image viewing systems remote from scene S and through communication link 740 and/or network 750 or through a wireless network such as 5G computer system 10 and application 624 through more user systems 720, 722, 724. Here, the computer system 10 output application 624 may receive two digital images of a plurality of manipulated scenes S and display the multi-dimensional image sequence 1010 to one or more user systems 720, 722, 724 via the communication link 740 and/or the network 750, or via a wireless network such as the 5G computer system 10 and the application 624.
In addition, the set of images (n) of scene S configured with respect to key body plane KSP may be transmitted as a multi-dimensional image sequence 1010 over communication link 740 and/or network 750, wireless, such as 5G, second computer system 10 and application 624 to display 628 to enable a plurality of users U to view multi-dimensional image sequence 1010 on display 628 live or as a replay/rebroadcast in a block or step 1250.
Referring now to FIG. 13, by way of example and not limitation, a touch screen display 628 is illustrated that enables user U to select a photography option of computer system 10. A first exemplary option may be DIFY capture, where user U may specify or select a digital image speed setting 1302, where user U may increase or decrease the playback speed or frames (images) per second of sequential display of digital images on display 628 multi-dimensional image/sequence 1010. Further, the user U may designate or select the number of loops or repetitions 1304 of the digital image to set the number of loops of the images (n) of the plurality of 2D images 1000 of the scene S, wherein the images (n) of the plurality of 2D images 1000 of the scene S are displayed in sequence on the display 628, similar to fig. 11. Still further, the user U may specify or select a playback order 1306 of the digital image sequence or the palindromic sequence for playback to set a display order of the images (n) of the multi-dimensional image/sequence 1010 of the scene S. The timing sequence of the images shows that the proper binocular disparity is produced by the motion chase ratio effect. It is contemplated herein that the computer system 10 and the application 624 may utilize default or automatic settings herein.
DIFY, referring to fig. 14A and 14B, by way of example and not limitation, there is illustrated a representation of frames captured in a set sequence, which are played back to the eye in the set sequence, and the perception of DIFY on human eye viewing display 628. DIFY and its geometry are interpreted to produce motion parallax. Motion parallax refers to the angular change of a point relative to a stationary point. (motion chase). Note that since we have set key subject KS points, all points of the foreground will move to the right, while all points of the background will move to the left. In the reverse palindromic of the image, the motion is reversed. In different views, any change in the angle of any point relative to the critical subject will produce motion parallax.
DIFY is a series of frames captured in a set sequence that is played back to the eye as a loop. For example, the playback case of two frames (assuming the first and last frames, such as frames 1101 and 1104) is depicted in fig. 14A. Fig. 14A shows the position of an object, such as the near plane NP object in fig. 4, on the near plane NP and its relationship to key subject KS points in frames 1101 and 1104, where the key subject KS points are constant due to the image translation imposed on frames 1101, 1102, 1103 and 1104. Frames 1101, 1102, 1103, and 1104 in fig. 11A and 11B may be overlapped and offset from the main axis 1112 by the calculated parallax value (horizontal image translation (HIT)), and preset by the pitch of the virtual cameras. Fig. 14B illustrates, by way of example and not limitation, a situation where the human eye perceives from viewing two frames depicted on display 628 of fig. 14A (assuming first and last frames, e.g., frames 1101 and 1104, with frame 1 in the near plane NP being point 1401 and frame 2 in the near plane NP being point 1402), wherein the image plane or screen plane is the same as key subject KS point and key subject plane KSP, user U viewing display 628 views virtual depth near plane NP 1410 in front of display 628 or between display 628 and user U eyes-left eye LE and right eye RE. The virtual depth near plane NP 1410 is the near plane NP because it represents the object with frame 1 in the near plane NP as near plane point 1401 and frame 2 in the near plane NP as near plane point 1402, which is the closest point the user U-eye-left eye LE and right eye RE see when viewing the multi-dimensional image sequence 1010 on the display 628.
Virtual depth near plane NP 1410 simulates the visual depth between key body KS and objects in near plane point 1401 and objects in near plane point 1402 as virtual depth 1420, i.e., the depth between near plane NP and key body plane KSP. This depth is due to binocular disparity between two views of the same point (object in near plane point 1401 and object in near plane point 1402). The objects in the near-plane point 1401 and the objects in the near-plane point 1402 are preferably the same points in the scene S at different views ordered in time due to binocular disparity. In addition, the external ray 1430 and more particularly the viewing angle 1440 of the user U's eyes-left eye LE and right eye RE-are preferably at about twenty-seven (27) degrees to the retina or ocular axis. (similar to the depth of field of a cell phone or tablet computer using display 628). This depiction helps define the boundaries of the composition of scene S, near-plane points 1401 and 1402 are preferably located within the depth of field, within outer ray 1430, while near plane NP must be outside of the inner intersection location 1450 of outer ray 1430.
The motion from X1 to X2 is the motion that the user U eye-left eye LE and right eye RE will track. Xn is the distance of the eye lens, left eye LE, or right eye RE to the image point 1411, 1412 on the virtual near image plane 1410. X' n is the distance from the right triangle of Xn to the formed route (leg) from the eye lens, left eye LE or right eye RE to the image points 1411, 1412 on the virtual near image plane 1410 to the image planes 628, KS, KSP. The smooth motion is binocular parallax caused by the offset relative to the key subject KS at each of the points observed by the user U eyes-left eye LE and right eye RE.
For each eye, either the left eye LE or the right eye RE, a coordinate system may be established relative to the eye center CL and half-1440 of the inter-ocular separation center-pupil distance width IPD. The two angles β and α are angles that are used to explain the dip motion pursuit. Beta is the angle formed as a line passes from the eye lens-left eye LE and right eye RE-through virtual near plane 1410 to the images on image planes 628, KS, KSP. Θ is β2- β1. And α is the angle of the fixed key body KS of the two frames 1101, 1104 on the image planes 628, KS, KSP to the points 1411, 1412 on the virtual approximation image plane 1410. The change in alpha represents a chase of the eye. The movement of the eyeball rotation follows the change in the position of a point on the virtual approximate plane. And β is the angle responsible for smooth motion or binocular parallax when the left and right eyes are compared. The external rays 1430 emitted from the eye lens, left eye LE and right eye RE, which are connected to the 1440 point, represent the depth of field or edges of the image, i.e. half of the image. The line will change as the depth of field of the virtual camera changes.
di/f=Xi
If we define the chase motion as the difference in position of points along the virtual near plane, we can derive by using tangent lines:
these equations indicate to us that the chase motion, X2-X1, is not a direct function of the line of sight. As the line of sight increases, the perceived depth di will become smaller, but due to the small angular difference, the motion will remain approximately the same with respect to the entire width of the image.
Mathematically, the ratio of retinal motion to eye smooth chasing rate determines the depth of a fixed point relative to the center of human vision. The creation of KSP provides the fixed point needed to create depth. Mathematically, all points will move in a different way than any other point, since the reference point is the same in all cases.
Referring now to FIG. 17, by way of example and not limitation, a representative illustration of a comfort circle (CoC) fused with a point of view (Horopter) arc or point and Panum region is shown. The point of gaze is a locus of points in space that have the same parallax as the point of gaze, the point of gaze arc or the point. Objects in the scene that fall near the point of gaze or point are sharp images, while those that are outside (in front of or behind) the point of gaze or point are blurred or blurred. Panum is a spatial region, panum region 1720, surrounding the same viewpoint to achieve a given degree of ocular convergence, inner boundary 1721, outer boundary 1722, where different points projected to the left and right eyes LE/RE result in fusion of the eyes, producing a perception of visual depth, while points outside this region result in a double view, a dual image. In addition, for objects falling within the Panum area, including objects near the same viewpoint, the images of the left and right eyes are fused and the user U will see a single sharp image. Outside the Panum's area, user U sees a dual image, either in front or behind.
It is recognized herein that computer system 10 may be executed by image capture application 624, image manipulation application 624, image display application 624 using different and separately located computer systems 10, such as one or more user systems 220, 222, 224, and application programs 206. Next, the second computer system 10 and application 206 may transmit the image set (n) of the scene S relative to the critical subject plane over the communication link 240 and/or network 250, wireless such as 5G, introducing (left and right) binocular disparity to display the multi-dimensional digital image on the display 628 to enable the plurality of users U to view the multi-dimensional digital image live or as rebroadcast/rebroadcast on the display 628 in block or step 1250.
Further, fig. 17 shows that the multi-dimensional image 1010 is displayed and viewed on the display 628 by left and right pixels 1550L/R, the light of the multi-dimensional image 1010 passing through lenticular lens 1540 and bending or refracting to provide 3D viewing of the multi-dimensional image 1010 on the display 628 to the left eye LE and right eye RE, viewing distance VD from pixel 1550, wherein the near object, key subject KS, and far object are within a comfort circle CoC, which is a near side point arc or point and within Panum region 1720 to achieve clear, single image 3D viewing of the multi-dimensional image 1010 on the display 628 that is comfortable and compatible with the human visual system of the user U.
Blockchains are a type of shared ledger that facilitates the process of logging transactions and tracking assets in a network. Assets may be tangible (property, car, cash, land, artwork or DIGY, 3D stereo and ownership of its data set) or intangible (intellectual property, patent, copyright, brand). Blockchains are a kind of decentralized, distributed and commonly disclosed digital ledgers, consisting of records called blocks that are used to record transactions on multiple computers using cryptography so that any block involved cannot be changed retrospectively without the need to change all subsequent blocks. This allows the participants to independently verify and audit the transaction. Each chunk contains the data or set of data, the cryptographic hash of the previous chunk (chain), the timestamp, and the cryptographic hash (identifying the chunk and all its contents). The timestamp proves that transaction data already exists when the block is published to obtain its hash. Since the blocks each contain information about their previous block, they form a chain and each additional block will augment its previous block. Thus, blockchains have difficulty modifying their data because once recorded, the data in any given block cannot be changed retrospectively without changing all subsequent blocks. If a block is changed, all subsequent blocks are invalid and have different hashes. Ensuring security through a "proof of job" mechanism slows down the transaction rate of creating new blocks and doing distributed processing. Blockchains are typically managed by a point-to-point network (i.e., an open and decentralized database) that acts as a publicly distributed ledger, where nodes collectively adhere to a protocol to communicate and authenticate new blocks. Each new tile is sent to everyone on the network to verify the new tile, thereby achieving consensus. Although blockchain records are not immutable, because forks are possible, blockchains may be considered secure via the design of a distributed computing system.
Ethernet is a popular blockchain standard and other proprietary and open source blockchain algorithms.
DIGY, 3D stereo, datasets and other datasets as well as any authentication documents, authenticity statements, histories, dates of origin, ownership chains, owner chains, etc. (authentications) that verify the creator or author may be contained in the Dataset (Dataset). The data set may be stored as chunks on the blockchain network, and thereafter each subsequent transaction related to the data set may be set forth in a subsequent chunk, each subsequent chunk containing a cryptographic hash of the previous chunk, the data set transaction data being a timestamp of the decentralized trusted source.
The non-replaceable token (NFT) is a unique collectable encrypted asset stored on the blockchain with a unique identification code and metadata that distinguishes the DIGY, DIGY sequence, or stereoscopic 3D image file, dataset, or dataset from any other digital content for the selected DIGY, DIGY sequence, or stereoscopic 3D image file, dataset, or other dataset, making it a unique one or limited version of the digital content. Unique digital content and/or a certain number of copies-unique or limited number of digital mementos that cannot be copied-can be created. Souvenirs are stored on a blockchain network. It is "irreplaceable" in that it cannot be exchanged as easily as similar replicable content that is freely moved through the internet. Furthermore, verifying precious artwork using DIGY, DIGY sequences or stereoscopic 3D image files, artwork data sets and evaluations, and any authentication files proves that an author or author, a statement of authenticity, history, date of origin, ownership chain, owner chain, etc. (authentication) may be included in the data set (Dataset), and an NFT of the smart contract or data set is made.
NFT simply records who owns the unique digital content. The content may be any content of art, music, photos, 3D graphics, tweets, memos, games, video, GIF, etc. It can become an NFT as long as it is digital content and as long as it is created.
Current NFT digital content may be captured by third party devices in piracy and disseminated in large amounts via the internet, and devaluating NFT digital content of standard NFT content.
The digital content here is DIGY and 3D stereoscopic images generated by artist users and creates own NFT artwork based on DIGY and 3D stereoscopic images using our above DIGY and 3D stereoscopic platforms.
Referring now to fig. 18, a flow chart 1800 of a method of creating an NFT for DIGY, DIGY sequences, stereoscopic 3D image files, or other data files (datasets) is shown.
In block or step 1810, an application NFT (application 206) on the smart device (computer system 10) represented by fig. 13 and 16B is opened, clicked or launched.
In block or step 1815, click/touch/select "create NFT" on the smart device.
In block or step 1820, the "DIGY/DIGY sequence/stereoscopic 3D image File library" icon on display 628 is clicked/touched/selected/accessed to access the image file or the like on the smart device.
In block or step 1825, DIGY sequence, stereoscopic 3D image file, or other data file (Dataset) file from the library master storage device 214 of the smart device 10 is clicked/touched/selected.
In block or step 1830, click/touch/select "create NFT" on display 628 of smart device 10.
In block or step 1835, an NFT is created for a selected image file, DIGY sequence, stereoscopic 3D image file, or other data file (Dataset). The user of the smart device 10 will issue NFT for its DIGY, DIGY sequence, stereoscopic 3D image file, or other data file (Dataset) using one of the blockchain providers via the system or application. Ethernet is the leading blockchain service for NFT distribution. However, there are a range of other blockchains that are becoming increasingly popular. One of which may be provided by the smart device manufacturer or other service provider.
In block or step 1840, NFT sales or exchanges of DIGY, DIGY sequences, stereoscopic 3D image files, or other data files (datasets) are provided using NFT token standards, compatible wallet services, and markets. Once created, the user may provide his NFT, such as NFT-DIGY, DIGY sequence, or stereoscopic 3D image file, for sale or verification on NFT-compatible wallet services and markets (such as OpenSea, which is an ethernet-based NFT market). Similar markets may be provided by smart device manufacturers or other service providers.
DIGY, DIGY sequences, stereoscopic 3D image files, or other data files (Dataset) are unique in that they cannot be captured by third party devices, such as other intelligent devices that view DIGY, DIGY sequences, or stereoscopic 3D, because such devices cannot access the original digital files and may not have a license for DIGY and 3D stereoscopic platforms. Current NFT digital content may be captured by third party devices in piracy and disseminated in large amounts via the internet, and devaluating NFT digital content of standard NFT content.
Here DIGY may include 2D video, 2D image collage, 3DDIGY.
The sequence is to concatenate multiple DIGYS together or loop through sequentially to create one story.
Referring now to FIG. 19, a flowchart 1900 of a method of linking or looping a plurality of DIGY-MP4 files in sequence or as loops, with or without audio files such as AAC (m 4 a) format, AIFF, apple Lossless, MP3, and WAV or other similar audio formats stored therein, is shown.
In block or step 1910, the application Photon3D (application 206) on the display 628 of the smart device (computer system 10) represented by fig. 13 and 16B is opened, clicked or launched.
In block or step 1915, click/touch/select "Create DIGY sequence" on the display of smart device 10.
In block or step 1920, the "DIGY library" icon on display 628 is clicked/touched/selected to access, for example, a DIGY MP4 file or the like on smart device 10.
In block or step 1925, the first DIGY file from DIGY library main storage 214 of smart device 10 is clicked/touched/dragged/selected to the timeline.
In block or step 1930, the second or more DIGY files from the DIGY library on the display of smart device 10 are clicked/touched/dragged/selected and placed in the order or desired order.
In block or step 1935, the DIGY file is trimmed/edited/cut or the duration (start/stop/speed) is set. DIGY is deployed by setting up transitions/cuts. An optional audio file/sound effect is added as the second track and the sound amplitude is mixed.
In block or step 1940, reordering the DIGY files by clicking/touching/selecting/dragging via dragging the DIGY files to a different order or position is sequential.
In block or step 1945, the completed DIGY sequence-MP 4 is saved by clicking/touching/selecting save to save the completed image and audio MP4 file in the main storage 214 of the smart device 10.
In block or step 1950, the completed DIGY sequence-MP 4 file is viewed and listened to on the smart device (computer system 10).
In block or step 1955, the shared DIGY sequence-MP 4 file may be attached to an email or text, airdrop, or uploaded to social media for sharing.
It is contemplated herein that a user can match the transition between two different DIGY, whether implemented manually or programmatically using audio peak detection or even a temporal feature/tempo matching algorithm BPM: 120 beats per minute; and (3) marking: 4/4.4 section = 4 segment (DIGY), with 3 transitions. [ DIGY 1-DIGY 2-DIGY 3-DIGY4].4 bars x 4 beats per bar = 16 beats; 16 beats @120bpm = 8 seconds; 4 segments (DIGY) @ 2 seconds per segment = 8 seconds. 1 digy=4 DIGY per segment.
It is contemplated herein that the user may use transient detection to find a beat down at the beginning of the bar and then synchronize it with the beginning of the first DIGY. Each DIGY is trimmed to 2 seconds and then ordered together to match the music transitions of the ordered DIGY.
Referring now to FIG. 20, a flowchart 2000 of a method of manipulating DIGY or DIGY sequence image files to synchronize with an audio file is shown.
In block or step 2010, an application Photon3D (application 206) on the display 628 of the smart device (computer system 10) represented by fig. 13 and 16B is opened, clicked or launched.
In block or step 2015, the "music application" is clicked/touched/selected via input to the display 628 of the smart device 10.
In block or step 2020, DIGY sequence, stereoscopic 3D image file (dataset) is clicked/touched/selected via input to the display 628 of the smart device 10 to select the dataset file from the storage means 604, 606 via input from said display 628.
In block or step 2025, instructions 206 are executed via processor 102 to prepare, converting the DIGY.gif file to a DIGY.mps file via input from the display 628; creating a DIGY frame timeline via input from display 628; importing image frames (3D igy, 2D image, 2D video) into the frame timeline via input from the display 206; the sequence/frame dwell time is adjusted via input from the display 206.
In block or step 2030, instructions 206 are executed via processor 102 to record an audio file via a microphone in smart device 10, select the audio file from storage 604, 606 or an online service such as ITUNES via input from the display 628, and import or download the audio file to a timeline via input from the display 628. It is contemplated herein that DIGY sequences and audio files may be converted to mp4 files for sharing with other smart devices 222 via network 250.
In block or step 2035, the instructions 206 are executed via the processor 102 to drag, overlay, or drop the audio file from the memory devices 604, 606 to the DIGY frame timeline via input from the display 628.
In block or step 2040, instructions 206 are executed via processor 102 to adjust, crop, link, or arrange the audio file relative to the DIGY image file via input from display 628.
In block or step 2045, the completed image and audio files-MP 4 are saved by clicking/touching/selecting save to save the completed image and audio MP4 files in main storage 214 via input from display 628.
In block or step 2050, the completed image and audio files are played, viewed and listened to on the smart device (computer system 10) via input from the display 628.
In block or step 2055, the shared DIGY and audio (files) may be attached to email or text, air drops, or uploaded to social media for sharing.
With respect to the above description, it should be appreciated that the optimum dimensional relationships, including variations in size, material, shape, form, position, movement mechanism, function and manner of operation, assembly and use, are intended to be covered by the present disclosure.
The foregoing description and drawings include exemplary embodiments. Having thus described exemplary embodiments, it will be noted by those skilled in the art that the present disclosure is exemplary only and that various other substitutions, modifications, and alterations may be made within the scope of the present disclosure. The mere listing or numbering of the steps of a method in a certain order does not constitute any limitation on the order of the steps of the method. Many modifications and other embodiments will come to mind to one skilled in the art to which this disclosure pertains having the benefit of the teachings presented in the foregoing descriptions and the associated drawings. Although specific terms may be employed herein, they are used in a generic and descriptive sense only and not for purposes of limitation. Further, the present disclosure has been described in detail, it should be understood that various changes, substitutions, and alterations can be made hereto without departing from the spirit and scope of the disclosure as defined by the appended claims. Accordingly, the present disclosure is not to be limited by the specific embodiments illustrated herein, but only by the appended claims.

Claims (23)

1. A system for simulating a 3D image from a pair of 2D images of a scene, the system comprising:
A smart device, the smart device having:
memory means for storing instructions;
a processor in communication with the memory device, the processor configured to execute the instructions;
a plurality of digital image capturing devices in communication with the processor and each configured to capture a digital image of the scene, the plurality of digital image capturing devices positioned approximately linearly in series within an approximate inter-pupillary distance width, wherein a first digital image capturing device is positioned proximate a first end of the inter-pupillary distance width, a second digital image capturing device is positioned proximate a second end of the inter-pupillary distance width, and any remaining digital image capturing devices of the plurality of digital image capturing devices are evenly spaced between the first end of the inter-pupillary distance width and the second end of the inter-pupillary distance width to capture a series of 2D images of the scene; and
a display in communication with the processor, the display configured to display the multi-dimensional digital image.
2. The system of claim 1, wherein the processor executes instructions to save the series of 2D images of the scene in the memory device.
3. The system of claim 2, wherein the processor executes instructions to select pairs of 2D images from the series of 2D images of the scene to produce a 3D stereoscopic image.
4. The system of claim 3, wherein the processor executes instructions to select an automatic key subject selection algorithm via input from the display, wherein the processor identifies key subject points in the pair of images of the scene, and each image of the pair of images of the scene is aligned with the key subject point by the processor, and all other points in the pair of images of the scene are shifted based on the spacing of the plurality of digital image capture devices to generate a modified pair of 2D images.
5. The system of claim 3, wherein the processor executes instructions to select a manual key subject selection algorithm via input from the display, wherein the processor enables a user to position an icon within the scene via input from the display to identify key subject points in the pair of images of the scene, and each image of the pair of images of the scene is aligned with the key subject point by the processor, and all other points of the pair of images of the scene are shifted based on a pitch of the plurality of digital image capture devices to generate a modified pair of 2D images.
6. The system of claim 3, wherein the processor executes instructions to select key subject points in the pair of 2D images of the scene via input from the display.
7. The system of claim 1, wherein the processor executes instructions to define two or more planes for each of the pair of images of the scene, wherein the two or more planes have different depth estimates.
8. The system of claim 7, wherein the processor executes instructions to identify a first proximal plane and a second distal plane within the pair of 2D images of the scene.
9. The system of claim 8, wherein the processor executes instructions to determine depth estimates for the first and second proximal planes within the pair of 2D images of the scene.
10. The system of claim 9, wherein the processor executes instructions to horizontally and vertically align the first proximal plane of each image frame of the pair of images and shift the second distal plane of each subsequent image frame of a sequence based on the depth estimation of the second distal plane of the pair of images of the scene to produce a second modified pair of 2D images.
11. The system of claim 8, wherein the first proximal plane and the second distal plane further comprise at least a foreground plane and a background plane.
12. The system of claim 10, wherein the processor executes instructions to inter-phase the second modified pair of 2D images into a multi-dimensional digital image.
13. The system of claim 12, wherein the processor executes instructions to save the multi-dimensional digital image to the memory.
14. The system of claim 13, wherein the processor executes instructions to record an audio file via a microphone in communication with the processor.
15. The system of claim 14, wherein the processor executes instructions to save the audio file to the memory.
16. The system of claim 15, wherein the processor executes instructions to select a multi-dimensional digital image from the memory via input from the display and display the multi-dimensional digital image on the display.
17. The system of claim 16, wherein the processor executes instructions to select an audio file from the memory via input from the display and superimpose the audio file on the multi-dimensional digital image on the display.
18. The system of claim 17, wherein the processor executes instructions to crop the audio file via input from the display to align with the multi-dimensional digital image.
19. The system of claim 18, wherein the processor executes instructions to save the audio file and the multi-dimensional digital image via input from the display.
20. The system of claim 19, wherein the processor executes instructions to play the audio file and display the multi-dimensional digital image via input from the display.
21. The system of claim 20, wherein the processor executes instructions to share the audio file and the multi-dimensional digital image with a second smart device via input from the display.
22. The system of claim 20, wherein the processor executes instructions to generate the audio file and the non-replaceable token of the multi-dimensional digital image with a second smart device via input from the display.
23. The system of claim 22, wherein the processor executes instructions to exchange the audio file and the non-replaceable token of the multi-dimensional digital image via input from the display with a wallet service and marketplace.
CN202280047753.0A 2021-06-07 2022-06-07 2D digital image capturing system and analog 3D digital image and sequence Pending CN117897951A (en)

Applications Claiming Priority (18)

Application Number Priority Date Filing Date Title
US202163197941P 2021-06-07 2021-06-07
US63/197,941 2021-06-07
US202163212025P 2021-06-17 2021-06-17
US63/212,025 2021-06-17
US17/355,906 US20210321077A1 (en) 2020-01-09 2021-06-23 2d digital image capture system and simulating 3d digital image sequence
US17/355,906 2021-06-23
US17/459,067 US20210392314A1 (en) 2020-01-09 2021-08-27 Vehicle terrain capture system and display of 3d digital image and 3d sequence
US17/459,067 2021-08-27
US17/511,490 2021-10-26
US17/511,490 US20220051427A1 (en) 2020-01-09 2021-10-26 Subsurface imaging and display of 3d digital image and 3d image sequence
US17/525,246 2021-11-12
US17/525,246 US20220078392A1 (en) 2020-01-09 2021-11-12 2d digital image capture system, frame speed, and simulating 3d digital image sequence
US29816461 2021-11-22
US29/816,462 2021-11-22
US29816462 2021-11-22
US29/816,461 2021-11-22
US17/834,023 US20220385807A1 (en) 2021-05-28 2022-06-07 2d digital image capture system and simulating 3d digital image and sequence
PCT/US2022/032524 WO2022261111A1 (en) 2021-06-07 2022-06-07 2d digital image capture system and simulating 3d digital image and sequence

Publications (1)

Publication Number Publication Date
CN117897951A true CN117897951A (en) 2024-04-16

Family

ID=84425456

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202280047753.0A Pending CN117897951A (en) 2021-06-07 2022-06-07 2D digital image capturing system and analog 3D digital image and sequence

Country Status (2)

Country Link
CN (1) CN117897951A (en)
WO (2) WO2022261111A1 (en)

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2111702A1 (en) * 1992-04-17 1993-10-28 Hiroshi Miyawaki Correcting the position of a filming apparatus relative to printing paper in 3d photographic printing
US20060023197A1 (en) * 2004-07-27 2006-02-02 Joel Andrew H Method and system for automated production of autostereoscopic and animated prints and transparencies from digital and non-digital media
US10848731B2 (en) * 2012-02-24 2020-11-24 Matterport, Inc. Capturing and aligning panoramic image and depth data
US10178247B2 (en) * 2015-01-30 2019-01-08 Jerry Nims Digital multi-dimensional image photon platform system and methods of use
US9992473B2 (en) * 2015-01-30 2018-06-05 Jerry Nims Digital multi-dimensional image photon platform system and methods of use
US10074012B2 (en) * 2016-06-17 2018-09-11 Dolby Laboratories Licensing Corporation Sound and video object tracking
JP6829015B2 (en) * 2016-06-23 2021-02-10 マクセル株式会社 Mobile terminal
US10917633B2 (en) * 2017-09-15 2021-02-09 Intel Corporation Generating stereoscopic light field panoramas using concentric viewing circles
CN109756723B (en) * 2018-12-14 2021-06-11 深圳前海达闼云端智能科技有限公司 Method and apparatus for acquiring image, storage medium and electronic device
EP3694208A1 (en) * 2019-02-05 2020-08-12 Jerry Nims A method and system for simulating a 3-dimensional image sequence
US20220078392A1 (en) * 2020-01-09 2022-03-10 JUC Holdings Limited 2d digital image capture system, frame speed, and simulating 3d digital image sequence

Also Published As

Publication number Publication date
WO2022261105A1 (en) 2022-12-15
WO2022261111A1 (en) 2022-12-15

Similar Documents

Publication Publication Date Title
Kwon et al. A study of layout, rendering, and interaction methods for immersive graph visualization
US20220385880A1 (en) 2d digital image capture system and simulating 3d digital image and sequence
Balram et al. Light‐field imaging and display systems
WO2018076661A1 (en) Three-dimensional display apparatus
US20120020548A1 (en) Method for Generating Images of Multi-Views
US20240155096A1 (en) 2d image capture system & display of 3d digital image
US20220078392A1 (en) 2d digital image capture system, frame speed, and simulating 3d digital image sequence
JP6462293B2 (en) Display apparatus and method for providing multi-viewpoint video
Eldes et al. Multi-view autostereoscopic projection display using rotating screen
US20240169489A1 (en) Virtual, augmented, and mixed reality systems and methods
US20210392314A1 (en) Vehicle terrain capture system and display of 3d digital image and 3d sequence
US11818325B2 (en) Blended mode three dimensional display systems and methods
CN102474634B (en) Model amendment $$$$ image is shown for 3-dimensional
US20210297647A1 (en) 2d image capture system, transmission & display of 3d digital image
US20210321077A1 (en) 2d digital image capture system and simulating 3d digital image sequence
CN117897951A (en) 2D digital image capturing system and analog 3D digital image and sequence
EP4233008A1 (en) Vehicle terrain capture system and display of 3d digital image and 3d sequence
CN101908233A (en) Method and system for producing plural viewpoint picture for three-dimensional image reconstruction
EP4352954A1 (en) 2d digital image capture system and simulating 3d digital image and sequence
US20240236285A1 (en) 2d image capture system & display of 3d digital image
CN116076071A (en) Two-dimensional image capturing system and display of three-dimensional digital image
CN116097167A (en) Two-dimensional image capturing system and transmission and display of three-dimensional digital images
EP4070153A1 (en) Virtual, augmented, and mixed reality systems and methods
CN116097644A (en) 2D digital image capturing system and analog 3D digital image sequence
CN117203668A (en) 2D digital image capturing system, frame rate and analog 3D digital image sequence

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination