WO2012165218A1 - 記録装置、記録方法、再生装置、再生方法、プログラム、および記録再生装置 - Google Patents
記録装置、記録方法、再生装置、再生方法、プログラム、および記録再生装置 Download PDFInfo
- Publication number
- WO2012165218A1 WO2012165218A1 PCT/JP2012/063031 JP2012063031W WO2012165218A1 WO 2012165218 A1 WO2012165218 A1 WO 2012165218A1 JP 2012063031 W JP2012063031 W JP 2012063031W WO 2012165218 A1 WO2012165218 A1 WO 2012165218A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- camera
- view video
- base view
- information
- video stream
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/77—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera
- H04N5/772—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television camera the recording apparatus and the television camera being placed in the same enclosure
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/189—Recording image signals; Reproducing recorded image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/178—Metadata, e.g. disparity information
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
- H04N5/92—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/8042—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
- H04N9/8227—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being at least another television signal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
Definitions
- the present technology particularly relates to a recording apparatus, a recording method, a reproducing apparatus, a reproducing method, a program, and a recording / reproducing apparatus that can provide information on a plurality of cameras used for photographing to a reproducing apparatus.
- the video data of 3D content includes data for an image for the left eye (L image) and an image for the right eye (R image). There is a shift corresponding to parallax between the subject appearing in the L image and the subject appearing in the R image.
- the subject can be recognized in three dimensions.
- the optimal parallax varies depending on the size of the display image on the display device, and the optimal parallax varies depending on the viewing distance.
- the present technology has been made in view of such a situation, and makes it possible to provide information on a plurality of cameras used for photographing to a playback-side apparatus.
- the recording apparatus encodes an image captured by the first camera and an image captured by the second camera using H.264 / MPEG-4 MVC, and generates a non-base view video stream.
- the SEI including at least one of information on the optical axis interval between the first camera and the second camera, information on convergence angle, and information on focal length is recorded as the SEI of each picture constituting
- a recording control unit for recording the Base-view video stream and the Non-base-view video stream in which the SEI is recorded on a recording medium.
- the first camera and the second camera each have at least a lens.
- An image sensor for the first camera and an image sensor for the second camera may be provided as the image sensor that performs photoelectric conversion of light taken into the lens, or one image sensor is the first.
- the camera and the second camera may be used in common.
- the first camera and the second camera may be provided inside the recording apparatus, or may be provided outside the recording apparatus. When provided outside the recording apparatus, images captured by the first camera and the second camera are supplied to the recording apparatus by wired or wireless communication.
- the first camera that captures an image for the left eye and the second camera that captures an image for the right eye can be further provided.
- the encoding unit encodes the left-eye image captured by the first camera as the Base view video stream, and the right-eye image captured by the second camera is the Non-image.
- -base view Can be encoded as a video stream.
- the encoding unit can record information including two values as the optical axis interval information.
- the optical axis interval between the first camera and the second camera is expressed by dividing one of the two values by the other value.
- the encoding unit can record information including two values as the convergence angle information.
- the convergence angle of the first camera and the second camera is expressed by dividing one of the two values by the other value.
- the playback device provides a Base obtained by encoding an image captured by the first camera and an image captured by the second camera using H.264 / MPEG-4CMVC.
- a view video stream, and an SEI including at least one of information on optical axis spacing, convergence angle information, and focal length information of the first camera and the second camera is recorded as an SEI of each picture
- a decoding unit that decodes the non-base video stream is recorded as an SEI of each picture.
- a display control unit configured to adjust and display a parallax between an image obtained by decoding the Base view video stream and an image obtained by decoding the Non-base view video stream based on the SEI; Can be provided.
- the display control unit displays an image obtained by decoding the Base-view video stream as a left-eye image, and an image obtained by decoding the Non-base view video stream is used as a right-eye image. Can be displayed.
- the recording / playback apparatus encodes an image shot by the first camera and an image shot by the second camera using H.264 / MPEG-4 MVC, and produces a non-base view video.
- the SEI of each picture constituting the stream the SEI including at least one of information on an optical axis interval between the first camera and the second camera, information on a convergence angle, and information on a focal length is used.
- information on a plurality of cameras used for shooting can be provided to a playback-side device.
- FIG. 1 is a diagram illustrating a configuration example of a recording / reproducing system according to an embodiment of the present technology.
- the recording / reproducing system of FIG. 1 includes a recording device 1, a reproducing device 2, and a display device 3.
- the playback device 2 and the display device 3 are connected via a cable 4 such as an HDMI (High Definition Multimedia Interface) cable.
- HDMI High Definition Multimedia Interface
- the recording device 1 is a video camera capable of capturing and recording 3D images.
- a lens 11R is provided at a right position from the recording apparatus 1 toward the subject, and a lens 11L is provided at a left position.
- a right-eye camera having an optical system for generating an R image based on the light captured by the lens 11R and an L image based on the light captured from the lens 11L are generated.
- a left-eye camera having an optical system is provided.
- the recording apparatus 1 encodes the R image and the L image by H.264 / MPEG-4 (MVC (Multi-view Video coding), and records them on an internal recording medium according to a recording format such as AVCHD (Advanced Video Codec High Definition). .
- MVC Multi-view Video coding
- AVCHD Advanced Video Codec High Definition
- H.264 / MPEG-4 MVC will be described with reference to FIG.
- a video stream called a Base view video stream and a video stream called a Non-base view video stream are defined.
- the Base view video stream is a stream obtained by encoding, for example, an L image of L images and R images by H.264 / AVC. As indicated by the vertical arrows in FIG. 2, predictive coding using a Base view picture as a reference image is not permitted.
- non-base view pictures are allowed to be predictive encoded using the Base view picture as a reference image.
- the data amount of the Non-base view video stream is the video of the L image. Compared to the data volume of the Base view video stream that is a stream.
- the encoding of the L image and the R image taken by the recording device 1 is performed by such H.264 / MPEG-4 MVC.
- the recording apparatus 1 when the L image and the R image are encoded, the information regarding the camera status of the left-eye camera used for capturing the L image and the right-eye camera used for capturing the R image is used. Camera information is recorded in the Non-base view video stream.
- the camera information includes information representing the following contents.
- Camera optical axis interval (base line length) [mm] 2.
- Angle of convergence [degree] 3.
- the base line length is a distance (mm) between the position PR, which is the optical axis position of the lens 11R, and the position PL, which is the optical axis position of the lens 11L.
- the convergence angle is an angle between a straight line connecting the position P and the position PR and a straight line connecting the position P and the position PL, where the position of the subject is the position P.
- the 35mm equivalent focal length is the focal length (mm) in terms of 35mm between the left-eye camera and the right-eye camera at the time of shooting.
- the recording device 1 defines a data structure called MDP (Modified DV pack Meta) in user_data_unregistered SEI of the non-base view video stream, and records camera information using the data structure.
- MDP Modified DV pack Meta
- user_data_unregistered SEI is arbitrary user data and is added to each picture.
- the MDP is additional information recorded in real time at the time of shooting.
- the playback device 2 captures and plays back 3D video data captured by the recording device 1 and recorded on a recording medium inside the recording device 1 via a USB (Universal Serial Bus) cable or an HDMI cable.
- USB Universal Serial Bus
- the playback device 2 records the parallax between the L image obtained by playing the Base view video stream and the R image obtained by playing the Non-base view video stream in the Non-base view video stream at the time of shooting. Adjust using the camera information. For adjusting the parallax, information such as the size of the display device of the display device 3 is also used as appropriate so that the parallax is optimal when displayed on the display device of the display device 3.
- the playback device 2 outputs the L image and the R image with adjusted parallax to the display device 3 via the cable 4 and displays them alternately.
- the display device 3 is a TV that supports the display of stereo images.
- the display device 3 is provided with a display device such as an LCD (Liquid Crystal Display).
- the recording device 1 reproduces the situation such as the installation position of the camera at the time of shooting.
- the device 2 can be provided.
- the playback device 2 can know the status of the camera at the time of shooting by referring to the camera information recorded in the Non-base view video stream. In addition, the playback device 2 can reproduce the positional relationship between the recording device 1 and the subject by adjusting the parallax between the L image and the R image according to the situation of the camera at the time of shooting. Can be provided.
- FIG. 4 is a block diagram illustrating a configuration example of the recording apparatus 1.
- the recording apparatus 1 includes a camera unit 21, a recording unit 22, and a recording medium 23.
- the camera unit 21 includes a right-eye camera 31R and a left-eye camera 31L.
- the right-eye camera 31R includes a lens 11R and an image sensor 12R, and performs photoelectric conversion of light captured by the lens 11R in the image sensor 12R.
- the right-eye camera 31R performs processing such as A / D conversion on a video signal obtained by performing photoelectric conversion, and outputs R image data.
- the left-eye camera 31L includes a lens 11L and an image sensor 12L, and performs photoelectric conversion of light captured by the lens 11L in the image sensor 12L.
- the left-eye camera 31L performs processing such as A / D conversion on a video signal obtained by performing photoelectric conversion, and outputs L image data.
- the recording unit 22 includes an MVC encoder 41, a camera information acquisition unit 42, and a recording control unit 43.
- the MVC encoder 41 encodes the R image captured by the right-eye camera 31R and the L image captured by the left-eye camera 31L using H.264 / MPEG-4 MVC.
- the MVC encoder 41 includes a Base view video encoder 51, a Non-base view video encoder 52, and an integration unit 53.
- the R image data output from the right-eye camera 31R is input to the Non-base video viewer 52
- the L image data output from the left-eye camera 31L is the Base video video encoder 51 and the Non-base video video encoder 52. Is input.
- the Base view video encoder 51 encodes the L image captured by the left-eye camera 31 ⁇ / b> L using H.264 / AVC, and outputs the Base view video stream to the integration unit 53.
- the non-base view video encoder 52 encodes the R image captured by the right-eye camera 31 ⁇ / b> R using the L image captured by the left-eye camera 31 ⁇ / b> L as a reference image, and generates a non-base view video stream. To do.
- the Non-base view video encoder 52 adds user_data_unregistered SEI representing the baseline length, the convergence angle, and the 35 mm equivalent focal length supplied from the camera information acquisition unit 42 as the SEI of each picture of the Non-base view video stream. . From the camera information acquisition unit 42, information on the base line length, the convergence angle, and the 35 mm equivalent focal length of the right-eye camera 31R and the left-eye camera 31L, which is camera information acquired from the camera unit 21, is supplied.
- FIG. 5 is a diagram showing the data structure of a Non-base view video stream.
- FIG. 5A is a diagram showing a data structure of an Access unit that stores data of a leading picture among pictures included in one GOP (Group Of Picture).
- one Access data unit stores data for one picture.
- the Access Unit storing the data of the first picture of the Non-base view is a View and dependency dependency representation delimiter, Subset and SPS (Sequence Parameter Set), PPS (Picture Parameter Set), SEI (Supplemental (Enhancement Information) and Slice.
- “View and dependency representation delimiter represents the beginning of Access Unit. Subset
- SPS contains the information regarding the encoding of the whole sequence
- PPS contains the information regarding the encoding of the picture in which data is stored in Access
- SEI is additional information and includes various SEI messages such as MVC_scalable_nesting SEI and user_data_unregistered SEI.
- user_data_unregistered SEI of parallax information (Offset metadata) of subtitles displayed at the time of reproduction is included.
- Information on the base line length, the convergence angle, and the 35 mm equivalent focal length acquired by the camera information acquisition unit 42 is recorded as user_data_unregistered SEI different from user_data_unregistered SEI of parallax information.
- Slice following SEI is the data of the first picture (R image) of 1 GOP. Subsequent to Slice, Filler Data, End of Sequence, and End of stream may be included.
- FIG. 5B is a diagram illustrating the data structure of the Access Unit that stores data of the second and subsequent pictures among pictures included in one GOP.
- the Access Unit storing the data of the second and subsequent pictures of 1 GOP is composed of a View and dependency dependency representation delimiter and a Slice.
- Information such as PPS and SEI may be included.
- the data structure of the Access unit of the Base view video stream generated by the Base view video encoder 51 is the data shown in FIG. 5 except that user_data_unregistered SEI indicating the base line length, the convergence angle, and the 35 mm equivalent focal length is not recorded. It has the same data structure as the structure.
- FIG. 6 is a diagram showing the syntax of user_data_unregistered SEI that stores MDP related to baseline length, convergence angle, and 35 mm equivalent focal length.
- the uuid_iso_iec_11578 on the second line is a 128-bit field and is set to “17ee8c60-f84d-11d9-8cd6-0800200c9a66”.
- the TypeIndicator on the 3rd line is a 32-bit field and represents the type of user data transmitted by the SEI message. “0x4D 44 50 4D” represents the MDP.
- ModifiedDVPackMeta () of the fifth and subsequent lines is set. ModifiedDVPackMeta () includes number_of_modified_dv_pack_entries and one_modified_dv_pack ().
- the number_of_modified_dv_pack_entries on the 6th line is an 8-bit field and represents the number of one_modified_dv_pack () included in the user_data_unregistered SEI.
- one_modified_dv_pack () includes mdp_id and mdp_data.
- Mdp_id on the 9th line is an 8-bit field and represents the type of one_modified_dv_pack () including this field.
- FIG. 7 is a diagram showing mdp_id. As shown in FIG. 7, the fact that mdp_id is 0x20 indicates that one_modified_dv_pack () is BASELINE LENGTH pack (). BASELINE LENGTH pack () is one_modified_dv_pack () including baseline length information as mdp_data.
- Mdp_id of 0x21 indicates that one_modified_dv_pack () is CONVERGENCE ANGLE pack ().
- CONVERGENCE ANGLE pack () is one_modified_dv_pack () including information on the convergence angle as mdp_data.
- the 35mm equivalent focal length is represented by FOCAL_LENGTH in the existing Consumer Camera2 Pack ().
- Consumer Camera2 Pack () is one_modified_dv_pack () whose mdp_id is 0x71.
- Mdp_data on the 10th line in FIG. 6 is a 32-bit field, and represents one of the base line length, the convergence angle, and the 35 mm equivalent focal length. A fixed value may always be set as mdp_data.
- ModifiedDVPackMeta The size of one ModifiedDVPackMeta () is 255 bytes or less including emulation prevention bytes. -Multiple pictures of user_data_unregistered SEI including ModifiedDVPackMeta () must not be added to one picture. -If there is no user_data_unregistered SEI message including ModifiedDVPackMeta () in the first field of the complementary field pair, the user_data_unregistered SEI message including ModifiedDVPackMeta () must not be included in the second field.
- the total of mdp_data is 255 bytes or less, and the total of user_data_unregistered_SEI is 511 bytes or less.
- FIG. 8 shows BASELINE LENGTH pack ().
- BASELINE LENGTH pack () includes mdp_id, mdp_data1, and mdp_data2.
- the mdp_id of BASELINE LENGTH pack () is 0x20 as described above.
- Each of mdp_data1 and mdp_data2 is a 16-bit field, and the baseline length in mm is represented by two values of mdp_data1 and mdp_data2.
- Baseline length (baseline length [mm]) is represented by the following formula (1).
- baseline length [mm] mdp_data1 / mdp_data2 (1)
- FIG. 9 is a diagram showing CONVERGENCE “ANGLE” pack ().
- CONVERGENCE ANGLE pack () includes mdp_id, mdp_data1, and mdp_data2.
- the mdp_id of CONVERGENCE ANGLE pack () is 0x21 as described above.
- mdp_data1 and mdp_data2 are 16-bit fields, respectively, and the convergence angle in degrees is represented by two values of mdp_data1 and mdp_data2.
- the Non-base video view video encoder 52 in FIG. 4 outputs the non-base video view video stream in which information on the base line length, the convergence angle, and the 35 mm equivalent focal length is recorded to the integration unit 53 in this way.
- the integration unit 53 integrates the Base view video stream supplied from the Base view video encoder 51 and the Non-base view video stream supplied from the Non-base view video encoder 52, and converts the H.264 / MPEG-4 MVC.
- the encoded data is output to the recording control unit 43.
- the camera information acquisition unit 42 acquires information on the baseline length, the convergence angle, and the 35 mm equivalent focal length from, for example, the camera unit 21 and outputs the information to the non-base video viewer 52.
- the recording control unit 43 records the encoded data supplied from the MVC encoder 41 on the recording medium 23 according to, for example, AVCHD.
- the recording medium 23 includes a flash memory, a hard disk, and the like, and records encoded data according to control by the recording control unit 43.
- a memory card inserted into a slot provided in the casing of the recording apparatus 1 may be used as the recording medium 23.
- the encoded data recorded on the recording medium 23 is transferred to the reproducing device 2 when the recording device 1 is connected to the reproducing device 2 by a USB cable or the like.
- step S1 the MVC encoder 41 encodes the image input from the camera unit 21 using H.264 / MPEG-4 MVC. That is, the Base view video encoder 51 encodes the L image captured by the left-eye camera 31L using H.264 / AVC, and generates a Base view video stream. The non-base-view video encoder 52 encodes the R image captured by the right-eye camera 31 ⁇ / b> R using the L image as a reference image as appropriate, and generates a non-base-view video stream.
- step S ⁇ b> 2 the camera information acquisition unit 42 acquires information on the current baseline length, convergence angle, and 35 mm equivalent focal length from the camera unit 21.
- the Non-base view video encoder 52 sets BASELINE LENGTH pack () representing the baseline length, CONVERGENCE ANGLE pack () representing the convergence angle, and 35 mm equivalent focal length to user_data_unregistered SEI of each picture of Non-base view.
- Record Consumer Camera2 Pack ().
- Base-view video stream generated by Base-view video encoder 51 and Non-base, generated by Non-base-view video encoder 52, and recorded BASELINE-LENGTH pack (), CONVERGENCE-ANGLE pack (), Consumer-Camera2 Pack ()
- the view video stream is integrated by the integration unit 53 and supplied to the recording control unit 43.
- step S 4 the recording control unit 43 records the H.264 / MPEG-4 MVC encoded data supplied from the integration unit 53 on the recording medium 23.
- step S5 the MVC encoder 41 determines whether or not an input image from the camera unit 21 exists. If it is determined that the input image exists, the process from step S1 is repeated. The processing in steps S1 to S4 is performed on the image data of each picture. On the other hand, if the camera unit 21 has finished shooting and it is determined in step S5 that no input image exists, the process is terminated.
- the recording apparatus 1 can record information on the base line length, the convergence angle, and the 35 mm equivalent focal length of the camera used for taking a stereo image in the encoded data of each picture of Non-base view.
- the camera information can be provided to the playback device 2.
- the recording device 1 does not record the camera information in the Base ⁇ ⁇ ⁇ view video stream, but records it in the Non-base view video stream, thereby reproducing the Base view video stream and displaying only the L image (2D image).
- the playback apparatus 2 can perform this. Since the Base view video stream is a stream encoded by H.264 / AVC, any video data captured from the recording device 1 can be used as long as the device is compatible with H.264 / AVC. Can be played.
- the baseline length information may be recorded with the convergence angle as a fixed value.
- the information on the convergence angle may be recorded with the base line length as a fixed value.
- only 35 mm equivalent focal length information may be recorded.
- FIG. 11 is a block diagram illustrating a configuration example of the playback device 2.
- the playback apparatus 2 includes an acquisition unit 101, a playback unit 102, and a display control unit 103.
- the acquisition unit 101 acquires 3D video data, which is H.264 / MPEG-4 MVC encoded data, from the recording device 1 via a USB cable or the like, and outputs it.
- the 3D video data output from the acquisition unit 101 is input to the Base view video decoder 121 and the Non-base view video decoder 122 of the MVC decoder 111.
- the 3D video data captured from the recording device 1 may be recorded once on a recording medium in the playback device 2, and the 3D video data may be acquired by the acquisition unit 101 from the recording medium.
- the playback unit 102 includes an MVC decoder 111 and a camera information extraction unit 112.
- the MVC decoder 111 decodes 3D video data by H.264 / MPEG-4 MVC.
- the MVC decoder 111 includes a Base view video decoder 121 and a Non-base view video decoder 122.
- the Base view video decoder 121 decodes the Base view video stream included in the 3D video supplied from the acquisition unit 101 by H.264 / AVC, and outputs an L image.
- the L image output from the Base view video decoder 121 is supplied to the non-base view video decoder 122 and the display control unit 103.
- the Non-base view video decoder 122 appropriately decodes the Non-base view video stream included in the 3D video supplied from the acquisition unit 101 using the L image decoded by the Base view video decoder 121 as a reference image, and R Output an image.
- the R image output from the non-base view video decoder 122 is supplied to the display control unit 103.
- the camera information extraction unit 112 uses BASELINE LENGTH pack () representing the base line length and CONVERGENCE ANGLE representing the convergence angle. pack (), Consumer Camera2 Pack () representing 35mm equivalent focal length is acquired.
- the camera information extraction unit 112 obtains the baseline length by performing the calculation of the above formula (1) based on the two values included in BASELINE LENGTH pack (), and based on the two values included in CONVERGENCE ANGLE pack (). Then, the convergence angle is obtained by calculating the above equation (2). In addition, the camera information extraction unit 112 identifies a 35 mm equivalent focal length from Consumer Camera2 Pack (). The camera information extraction unit 112 outputs information on the base line length, the convergence angle, and the 35 mm equivalent focal length to the display control unit 103.
- the display control unit 103 calculates the parallax between the L image supplied from the Base view video decoder 121 and the R image supplied from the Non-base view video decoder 122, the base line length and the convergence angle acquired by the camera information extraction unit 112. Adjust using the 35mm equivalent focal length information.
- the display control unit 103 communicates with the display device 3 to acquire information such as the size of the display device of the display device 3 from the display device 3 and use it for parallax adjustment.
- the display control unit 103 outputs the L image and the R image in which the parallax is adjusted so that the parallax is optimal when displayed on the display device of the display device 3 to the display device 3 for display.
- step S21 the MVC decoder 111 decodes the 3D video data input from the acquisition unit 101 using H.264 / MPEG-4 MVC. That is, the Base view video decoder 121 decodes the Base view video stream using H.264 / AVC. Further, the Non-base view video decoder 122 decodes the Non-base view video stream using the L image decoded by the Base view video decoder 121 as a reference image as appropriate.
- step S22 the camera information extracting unit 112 extracts BASELINE LENGTH pack (), CONVERGENCE ANGLE pack (), Consumer Camera2 pack () from user_data_unregistered SEI of each picture of Non-base view.
- the camera information extraction unit 112 controls display of base line length obtained based on BASELINE LENGTH pack (), convergence angle obtained based on CONVERGENCE ANGLE pack (), and 35mm equivalent focal length information obtained from Consumer Camera2 Pack (). Output to the unit 103.
- step S23 the display control unit 103 performs parallax adjustment processing.
- the parallax adjustment process will be described later with reference to the flowchart of FIG.
- step S24 the display control unit 103 outputs the L image and the R image adjusted for parallax to the display device 3 to display them.
- step S25 the MVC decoder 111 determines whether or not there is encoded data to be decoded. If it is determined that the encoded data is present, the process from step S21 is repeated. The processes in steps S21 to S24 are performed on the encoded data of each picture. On the other hand, since the decoding of all the encoded data has been completed, when it is determined in step S25 that no encoded data exists, the process is ended.
- step S31 the display control unit 103 communicates with the display device 3 and acquires the size of the display device of the display device 3.
- the size of the display device of the display device 3 may be set directly by the user for the playback device 2.
- step S32 the display control unit 103 acquires the viewing distance, which is the distance from the display device surface of the display device 3 to the user, from the display device 3.
- the viewing distance may be directly set by the user with respect to the playback device 2.
- the amount of parallax perceived by the user varies depending on the viewing distance. Therefore, when the viewing distance can be acquired, the parallax can be adjusted more precisely by using the information. In the following, a case where the parallax is adjusted without considering the viewing distance will be described.
- step S33 the display control unit 103 obtains the parallax between the L image and the R image on the display device of the display device 3 when the parallax is not adjusted.
- step S34 the display control unit 103 determines whether or not the parallax between the L image and the R image on the display device exceeds a threshold value.
- step S35 the display control unit 103 uses the L image and the R image as they are without adjusting the parallax.
- the data is output to the display device 3 and displayed.
- step S34 when it is determined in step S34 that the parallax between the L image and the R image on the display device exceeds the threshold value, in step S36, the display control unit 103 reduces the parallax between the L image and the R image. The L image and the R image are shifted in the horizontal direction on the display device and displayed. Thereafter, the process returns to step S23 in FIG. 12, and the subsequent processing is performed.
- FIG. 14 is a diagram for explaining the relationship between the convergence angle, the baseline length, and the parallax.
- the range indicated by the dashed arrow # 1 is a range captured by the left-eye camera 31L
- the range indicated by the dashed arrow # 2 is a range captured by the right-eye camera 31R.
- ⁇ represents the convergence angle
- B represents the baseline length.
- F represents the subject distance.
- the display control unit 103 When calculating the parallax on the display device, the display control unit 103 first shoots the image of the focus position (subject image) by shifting the horizontal direction by the left-eye camera 31L and the right-eye camera 31R. Find out if it is broken. The difference in the horizontal shift amount indicated by the solid arrow # 3 is the parallax X.
- the parallax amount on the display device can be calculated from the horizontal width information of the display device. It becomes possible to calculate (step S33).
- the display control unit 103 shifts the L image and the R image in the horizontal direction so that the parallax is reduced (step S36).
- the horizontal shooting range D [m] is obtained from the horizontal angle of view q and the subject distance F [m]. Assuming that the shooting range D [m] is shot at a resolution of 1920 pixels, the parallax size converted to pixels is 1920 ⁇ (B / D) [pixels].
- the parallax B is the difference between the shooting range D [m] of the left-eye camera 31L and the shooting range D [m] of the right-eye camera 31R.
- the 35 mm equivalent focal length and subject distance are already defined in AVCHD's Modified DV pack meta, and can be used.
- the 35 mm equivalent focal length is expressed by FOCAL_LENGTH as described above, and the subject distance (focus position) is expressed by FOCUS.
- the horizontal size of the display device of the display device 3 is S [m]
- the length corresponding to the parallax on the display device is S ⁇ (B / D) [m].
- the L image and the R image can be displayed with appropriate parallax by shifting the image in the horizontal direction to reduce parallax.
- the playback device 2 can display the L image and the R image with an optimal parallax.
- FIG. 16 is a block diagram illustrating a configuration example of a recording / reproducing system.
- the recording device 1 in FIG. 16 includes a reproduction unit 102 and a display control unit 103. That is, the recording apparatus 1 in FIG. 16 has not only a function as a recording apparatus but also a function as a reproducing apparatus.
- 3D video is shot and 3D video obtained by shooting is played back as described with reference to FIG.
- the L image and the R image reproduced by the recording device 1 are output to the display device 3 and displayed.
- the recording device 1 and the display device 3 are connected by, for example, an HDMI cable.
- FIG. 17 is a block diagram showing another configuration example of the recording / reproducing system. 17 includes a camera unit 21, a recording unit 22, and a recording medium 23, and the playback device 2 includes an acquisition unit 101 and a playback unit 102.
- the display device 3 is provided with a display control unit 103.
- the playback device 2 in FIG. 17 has a function of decoding 3D video captured from the recording device 1 by H.264 / MPEG-4 MVC. However, unlike the playback device 2 in FIG. It does not have a function of adjusting the parallax between the obtained L image and R image. Adjustment of the parallax between the L image and the R image is performed in the display device 3.
- the acquisition unit 101 of the playback device 2 acquires 3D video data from the recording device 1, and the playback unit 102 decodes the 3D video data acquired by the acquisition unit 101 using H.264 / MPEG-4 MVC.
- the playback unit 102 outputs the L image and the R image obtained by decoding to the display device 3 together with information on the base line length, the convergence angle, and the 35 mm equivalent focal length acquired from user_data_unregistered SEI of the non-base video stream.
- the display control unit 103 of the display device 3 displays the parallax between the L image and the R image supplied from the playback device 2 based on the base line length, the convergence angle, and the 35 mm equivalent focal length information supplied from the playback device 2.
- the L image and the R image adjusted as described with reference to FIG. 13 and adjusted in parallax are displayed.
- information such as the size of the display device used for parallax adjustment is information known to the display device 3 that performs parallax adjustment.
- the camera information of the camera used to shoot the L image and the camera used to shoot the R image is recorded in the Non-base view video stream of H.264 / MPEG-4 MVC. It is also possible to record the camera information in one stream in which the L image and the R image are recorded by the Side by side method.
- H.264 AVC is used for the encoding method of L image and R image of Side by side method.
- the camera information is recorded, for example, in user_data_unregistered_SEI () of the AVC stream.
- FIG. 18 is a block diagram illustrating a configuration example of hardware of a computer that executes the above-described series of processing by a program.
- a CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- An input / output interface 155 is further connected to the bus 154.
- the input / output interface 155 is connected to an input unit 156 composed of a keyboard, a mouse, etc., and an output unit 157 composed of a display, a speaker, and the like.
- the input / output interface 155 is connected to a storage unit 158 made up of a hard disk, a non-volatile memory, etc., a communication unit 159 made up of a network interface, etc., and a drive 160 that drives the removable media 161.
- the CPU 151 loads the program stored in the storage unit 158 to the RAM 153 via the input / output interface 155 and the bus 154 and executes it, thereby executing the above-described series of processing. Is done.
- the program executed by the CPU 151 is recorded on the removable medium 161 or provided via a wired or wireless transmission medium such as a local area network, the Internet, or digital broadcasting, and is installed in the storage unit 158.
- the program executed by the computer may be a program that is processed in time series in the order described in this specification, or in parallel or at a necessary timing such as when a call is made. It may be a program for processing.
- Embodiments of the present technology are not limited to the above-described embodiments, and various modifications can be made without departing from the gist of the present technology.
- the first camera and the second camera are encoded by H.264 / MPEG-4 MVC, and the SEI of each picture constituting the Non-base view video stream is used as the first SEI.
- a recording apparatus comprising: a recording control unit configured to record a Base view video stream and a Non-base view video stream in which the SEI is recorded on a recording medium.
- the encoding unit includes: Encode the image for the left eye taken by the first camera as the Base view video stream,
- the encoding unit records information including two values as information on the optical axis interval,
- the optical axis interval between the first camera and the second camera is expressed by dividing one of the two values by the other value.
- the encoding unit records information including two values as the convergence angle information, The convergence angle of the first camera and the second camera is expressed by dividing one of the two values by the other value. Any one of (1) to (3) The recording device described.
- the image taken with the first camera and the image taken with the second camera are encoded by H.264 / MPEG-4 MVC,
- SEI of each picture constituting the non-base view video stream at least one of information on the optical axis interval between the first camera and the second camera, information on the convergence angle, and information on the focal length
- Record the SEI including A recording method including a step of recording a Base view video stream and a Non-base view video stream in which the SEI is recorded on a recording medium.
- the image taken with the first camera and the image taken with the second camera are encoded by H.264 / MPEG-4 MVC,
- SEI of each picture constituting the non-base view video stream at least one of information on the optical axis interval between the first camera and the second camera, information on the convergence angle, and information on the focal length
- Record the SEI including A program that causes a computer to execute processing including a step of recording a Base view video stream and a Non-base view video stream in which the SEI is recorded on a recording medium.
- a Base view video stream obtained by encoding an image captured by the first camera and an image captured by the second camera with H.264 / MPEG-4 MVC, the first camera, The non-base view video stream in which the SEI including at least one of the optical axis interval information, the convergence angle information, and the focal length information of the second camera is recorded as the SEI of each picture is decoded.
- a playback device including a decoding unit.
- a display control unit configured to adjust and display a parallax between an image obtained by decoding the Base view video stream and an image obtained by decoding the Non-base view video stream based on the SEI The playback device according to (7).
- the display control unit Display the image obtained by decoding the Base view video stream as an image for the left eye,
- a reproduction method including steps.
- a program that causes a computer to execute processing including steps.
- the first camera and the second camera are encoded by H.264 / MPEG-4 MVC, and the SEI of each picture constituting the Non-base view video stream is used as the first SEI.
- An encoding unit that records the SEI including at least one of information on an optical axis interval between the first camera and the second camera, information on an angle of convergence, and information on a focal length;
- a recording control unit that records a Base view video stream and a Non-base view video stream in which the SEI is recorded on a recording medium, and the Base view video stream and the Non-base view video stream recorded on the recording medium.
- a recording / reproducing apparatus that records a Base view video stream and a Non-base view video stream in which the SEI is recorded on a recording medium, and the Base view video stream and the Non-base view video stream recorded on the recording medium.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Library & Information Science (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
1.記録再生システムについて
2.記録装置について
3.再生装置について
4.変形例
図1は、本技術の一実施形態に係る記録再生システムの構成例を示す図である。
1. カメラ光軸間隔(基線長)[mm]
2. 輻輳角[度]
3. 35mm換算焦点距離[mm]
[記録装置の構成]
図4は、記録装置1の構成例を示すブロック図である。記録装置1は、カメラ部21、記録部22、および記録媒体23から構成される。
・emulation prevention bytesを含めて、1つのModifiedDVPackMeta()のサイズは255 bytes以下である。
・1ピクチャには、ModifiedDVPackMeta()を含む複数のuser_data_unregistered SEIメッセージを付加してはいけない。
・complementary fieldペアの最初のフィールドにModifiedDVPackMeta()を含むuser_data_unregistered SEIメッセージがない場合、2番目のフィールドにもModifiedDVPackMeta()を含むuser_data_unregistered SEIメッセージを含めてはいけない。
baseline length [mm] = mdp_data1/mdp_data2 ・・・ (1)
convergence angle [degree] = mdp_data1/mdp_data2 ・・・ (2)
ここで、図10のフローチャートを参照して、記録装置1の記録処理について説明する。図10の処理は、右目用カメラ31Rにより撮影されたR画像と左目用カメラ31Lにより撮影されたL画像が記録部22に入力されたときに開始される。
[再生装置の構成]
以上のようにして記録装置1により撮影され、記録媒体23に記録されたH.264/MPEG-4 MVCの符号化データを再生する再生装置2について説明する。
ここで、図12のフローチャートを参照して、再生装置2の再生処理について説明する。図12の処理は、例えば、3Dビデオのデータが取得部101から再生部102に入力されたときに開始される。
tan(q/2) = x/2f ・・・ (3)
変形例1.
図16は、記録再生システムの構成例を示すブロック図である。図16の記録装置1には、カメラ部21、記録部22、記録媒体23の他に、再生部102と表示制御部103が設けられる。すなわち、図16の記録装置1は、記録装置としての機能だけでなく、再生装置としての機能をも有している。
図17は、記録再生システムの他の構成例を示すブロック図である。図17の記録装置1には、カメラ部21、記録部22、記録媒体23が設けられ、再生装置2には取得部101と再生部102が設けられる。表示装置3には表示制御部103が設けられる。
以上においては、L画像の撮影に用いられたカメラとR画像の撮影に用いられたカメラのカメラ情報をH.264/MPEG-4 MVCのNon-base viewビデオストリームに記録するものとしたが、カメラ情報を、Side by side方式でL画像とR画像を記録した1つのストリームに記録することも可能である。
上述した一連の処理は、ハードウェアにより実行することもできるし、ソフトウェアにより実行することもできる。一連の処理をソフトウェアにより実行する場合には、そのソフトウェアを構成するプログラムが、専用のハードウェアに組み込まれているコンピュータ、または汎用のパーソナルコンピュータなどに、プログラム記録媒体からインストールされる。
本技術は、以下のような構成をとることもできる。
第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化し、Non-base viewビデオストリームを構成する各ピクチャのSEIとして、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含む前記SEIを記録する符号化部と、
Base viewビデオストリームと、前記SEIが記録されたNon-base viewビデオストリームとを記録媒体に記録させる記録制御部と
を備える記録装置。
左目用の画像を撮影する前記第1のカメラと、
右目用の画像を撮影する前記第2のカメラと
をさらに備え、
前記符号化部は、
前記第1のカメラで撮影された左目用の画像を前記Base viewビデオストリームとして符号化し、
前記第2のカメラで撮影された右目用の画像を前記Non-base viewビデオストリームとして符号化する
前記(1)に記載の記録装置。
前記符号化部は、前記光軸間隔の情報として、2つの値を含む情報を記録し、
前記第1のカメラと前記第2のカメラの光軸間隔は、前記2つの値のうちの一方の値を他方の値で除算することによって表される
前記(1)または(2)に記載の記録装置。
前記符号化部は、前記輻輳角の情報として、2つの値を含む情報を記録し、
前記第1のカメラと前記第2のカメラの輻輳角は、前記2つの値のうちの一方の値を他方の値で除算することによって表される
前記(1)乃至(3)のいずれかに記載の記録装置。
第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化し、
Non-base viewビデオストリームを構成する各ピクチャのSEIとして、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含む前記SEIを記録し、
Base viewビデオストリームと、前記SEIが記録されたNon-base viewビデオストリームとを記録媒体に記録させる
ステップを含む記録方法。
第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化し、
Non-base viewビデオストリームを構成する各ピクチャのSEIとして、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含む前記SEIを記録し、
Base viewビデオストリームと、前記SEIが記録されたNon-base viewビデオストリームとを記録媒体に記録させる
ステップを含む処理をコンピュータに実行させるプログラム。
第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化することによって得られたBase viewビデオストリームと、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含むSEIが各ピクチャのSEIとして記録されたNon-base viewビデオストリームとを復号する復号部を備える
再生装置。
前記SEIに基づいて、前記Base viewビデオストリームを復号して得られた画像と、前記Non-base viewビデオストリームを復号して得られた画像との視差を調整し、表示させる表示制御部をさらに備える
前記(7)に記載の再生装置。
前記表示制御部は、
前記Base viewビデオストリームを復号して得られた画像を左目用の画像として表示させ、
前記Non-base viewビデオストリームを復号して得られた画像を右目用の画像として表示させる
前記(7)または(8)に記載の再生装置。
第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化することによって得られたBase viewビデオストリームと、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含むSEIが各ピクチャのSEIとして記録されたNon-base viewビデオストリームとを復号する
ステップを含む再生方法。
第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化することによって得られたBase viewビデオストリームと、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含むSEIが各ピクチャのSEIとして記録されたNon-base viewビデオストリームとを復号する
ステップを含む処理をコンピュータに実行させるプログラム。
第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化し、Non-base viewビデオストリームを構成する各ピクチャのSEIとして、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含む前記SEIを記録する符号化部と、
Base viewビデオストリームと、前記SEIが記録されたNon-base viewビデオストリームとを記録媒体に記録させる記録制御部と
前記記録媒体に記録された前記Base viewビデオストリームと前記Non-base viewビデオストリームとを復号する復号部と
を備える記録再生装置。
Claims (12)
- 第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化し、Non-base viewビデオストリームを構成する各ピクチャのSEIとして、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含む前記SEIを記録する符号化部と、
Base viewビデオストリームと、前記SEIが記録されたNon-base viewビデオストリームとを記録媒体に記録させる記録制御部と
を備える記録装置。 - 左目用の画像を撮影する前記第1のカメラと、
右目用の画像を撮影する前記第2のカメラと
をさらに備え、
前記符号化部は、
前記第1のカメラで撮影された左目用の画像を前記Base viewビデオストリームとして符号化し、
前記第2のカメラで撮影された右目用の画像を前記Non-base viewビデオストリームとして符号化する
請求項1に記載の記録装置。 - 前記符号化部は、前記光軸間隔の情報として、2つの値を含む情報を記録し、
前記第1のカメラと前記第2のカメラの光軸間隔は、前記2つの値のうちの一方の値を他方の値で除算することによって表される
請求項1に記載の記録装置。 - 前記符号化部は、前記輻輳角の情報として、2つの値を含む情報を記録し、
前記第1のカメラと前記第2のカメラの輻輳角は、前記2つの値のうちの一方の値を他方の値で除算することによって表される
請求項1に記載の記録装置。 - 第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化し、
Non-base viewビデオストリームを構成する各ピクチャのSEIとして、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含む前記SEIを記録し、
Base viewビデオストリームと、前記SEIが記録されたNon-base viewビデオストリームとを記録媒体に記録させる
ステップを含む記録方法。 - 第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化し、
Non-base viewビデオストリームを構成する各ピクチャのSEIとして、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含む前記SEIを記録し、
Base viewビデオストリームと、前記SEIが記録されたNon-base viewビデオストリームとを記録媒体に記録させる
ステップを含む処理をコンピュータに実行させるプログラム。 - 第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化することによって得られたBase viewビデオストリームと、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含むSEIが各ピクチャのSEIとして記録されたNon-base viewビデオストリームとを復号する復号部を備える
再生装置。 - 前記SEIに基づいて、前記Base viewビデオストリームを復号して得られた画像と、前記Non-base viewビデオストリームを復号して得られた画像との視差を調整し、表示させる表示制御部をさらに備える
請求項7に記載の再生装置。 - 前記表示制御部は、
前記Base viewビデオストリームを復号して得られた画像を左目用の画像として表示させ、
前記Non-base viewビデオストリームを復号して得られた画像を右目用の画像として表示させる
請求項8に記載の再生装置。 - 第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化することによって得られたBase viewビデオストリームと、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含むSEIが各ピクチャのSEIとして記録されたNon-base viewビデオストリームとを復号する
ステップを含む再生方法。 - 第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化することによって得られたBase viewビデオストリームと、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含むSEIが各ピクチャのSEIとして記録されたNon-base viewビデオストリームとを復号する
ステップを含む処理をコンピュータに実行させるプログラム。 - 第1のカメラで撮影された画像と第2のカメラで撮影された画像をH.264/MPEG-4 MVCによって符号化し、Non-base viewビデオストリームを構成する各ピクチャのSEIとして、前記第1のカメラと前記第2のカメラの光軸間隔の情報、輻輳角の情報、および、焦点距離の情報のうちの少なくともいずれかを含む前記SEIを記録する符号化部と、
Base viewビデオストリームと、前記SEIが記録されたNon-base viewビデオストリームとを記録媒体に記録させる記録制御部と、
前記記録媒体に記録された前記Base viewビデオストリームと前記Non-base viewビデオストリームとを復号する復号部と
を備える記録再生装置。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP12793581.5A EP2688303A4 (en) | 2011-05-30 | 2012-05-22 | RECORDING DEVICE, RECORDING METHOD, PLAYING DEVICE, PLAYING METHOD, PROGRAM AND RECORDING / REPLAYING DEVICE |
KR20137030803A KR20140030202A (ko) | 2011-05-30 | 2012-05-22 | 기록 장치, 기록 방법, 재생 장치, 재생 방법, 프로그램 및 기록 재생 장치 |
US14/118,081 US20140072271A1 (en) | 2011-05-30 | 2012-05-22 | Recording apparatus, recording method, reproduction apparatus, reproduction method, program, and recording reproduction apparatus |
CN201280024910.2A CN103548345A (zh) | 2011-05-30 | 2012-05-22 | 记录设备和方法、再现设备和方法、程序和记录再现设备 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2011120187A JP2012249137A (ja) | 2011-05-30 | 2011-05-30 | 記録装置、記録方法、再生装置、再生方法、プログラム、および記録再生装置 |
JP2011-120187 | 2011-05-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012165218A1 true WO2012165218A1 (ja) | 2012-12-06 |
Family
ID=47259079
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2012/063031 WO2012165218A1 (ja) | 2011-05-30 | 2012-05-22 | 記録装置、記録方法、再生装置、再生方法、プログラム、および記録再生装置 |
Country Status (7)
Country | Link |
---|---|
US (1) | US20140072271A1 (ja) |
EP (1) | EP2688303A4 (ja) |
JP (1) | JP2012249137A (ja) |
KR (1) | KR20140030202A (ja) |
CN (1) | CN103548345A (ja) |
TW (1) | TW201304540A (ja) |
WO (1) | WO2012165218A1 (ja) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10880565B2 (en) * | 2014-03-24 | 2020-12-29 | Qualcomm Incorporated | Use of specific HEVC SEI messages for multi-layer video codecs |
JP2015198391A (ja) * | 2014-04-02 | 2015-11-09 | キヤノン株式会社 | 撮像装置、撮像装置の制御方法、およびプログラム |
US10104415B2 (en) * | 2015-01-21 | 2018-10-16 | Microsoft Technology Licensing, Llc | Shared scene mesh data synchronisation |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007280516A (ja) | 2006-04-06 | 2007-10-25 | Sony Corp | 記録装置、記録方法および記録プログラム、記録再生装置、記録再生方法および記録再生プログラム、編集装置、編集方法および編集プログラム、ならびに、表示方法 |
JP2010157826A (ja) * | 2008-12-26 | 2010-07-15 | Victor Co Of Japan Ltd | 画像復号装置、画像符復号方法およびそのプログラム |
WO2010085361A2 (en) * | 2009-01-26 | 2010-07-29 | Thomson Licensing | Frame packing for video coding |
JP2011502375A (ja) * | 2007-10-10 | 2011-01-20 | 韓國電子通信研究院 | ステレオスコピックデータの保存および再生のためのメタデータ構造ならびにこれを利用するステレオスコピックコンテンツファイルの保存方法 |
WO2011024373A1 (ja) * | 2009-08-31 | 2011-03-03 | パナソニック株式会社 | 立体視制御装置、集積回路、立体視制御方法 |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE69030911T2 (de) * | 1989-10-25 | 1997-11-06 | Hitachi Ltd | Stereoskopisches bilderzeugendes System |
US7515759B2 (en) * | 2004-07-14 | 2009-04-07 | Sharp Laboratories Of America, Inc. | 3D video coding using sub-sequences |
KR101199498B1 (ko) * | 2005-03-31 | 2012-11-09 | 삼성전자주식회사 | 카메라 파라미터를 이용한 다시점 영상의 부호화 장치 및생성 장치, 그 방법과, 이를 수행하기 위한 프로그램이기록된 기록 매체 |
CN101292538B (zh) * | 2005-10-19 | 2012-11-28 | 汤姆森特许公司 | 使用可缩放的视频编码的多视图视频编码 |
EP2088789A3 (en) * | 2008-02-05 | 2012-08-15 | Samsung Electronics Co., Ltd. | Apparatus and method for generating and displaying media files |
KR101530713B1 (ko) * | 2008-02-05 | 2015-06-23 | 삼성전자주식회사 | 영상 파일을 생성하고 표시하기 위한 장치 및 방법 |
KR101506219B1 (ko) * | 2008-03-25 | 2015-03-27 | 삼성전자주식회사 | 3차원 영상 컨텐츠 제공 방법, 재생 방법, 그 장치 및 그기록매체 |
US9288470B2 (en) * | 2008-12-02 | 2016-03-15 | Lg Electronics Inc. | 3D image signal transmission method, 3D image display apparatus and signal processing method therein |
JP4957823B2 (ja) * | 2009-04-08 | 2012-06-20 | ソニー株式会社 | 再生装置および再生方法 |
US9124874B2 (en) * | 2009-06-05 | 2015-09-01 | Qualcomm Incorporated | Encoding of three-dimensional conversion information with two-dimensional video sequence |
KR20110011000A (ko) * | 2009-07-27 | 2011-02-08 | 삼성전자주식회사 | 3차원 영상 재생을 위한 부가 정보가 삽입된 3차원 영상 데이터스트림 생성 방법 및 그 장치, 3차원 영상 재생을 위한 부가 정보가 삽입된 3차원 영상 데이터스트림 수신 방법 및 그 장치 |
-
2011
- 2011-05-30 JP JP2011120187A patent/JP2012249137A/ja not_active Abandoned
-
2012
- 2012-05-22 KR KR20137030803A patent/KR20140030202A/ko not_active Application Discontinuation
- 2012-05-22 EP EP12793581.5A patent/EP2688303A4/en not_active Withdrawn
- 2012-05-22 US US14/118,081 patent/US20140072271A1/en not_active Abandoned
- 2012-05-22 TW TW101118189A patent/TW201304540A/zh unknown
- 2012-05-22 CN CN201280024910.2A patent/CN103548345A/zh active Pending
- 2012-05-22 WO PCT/JP2012/063031 patent/WO2012165218A1/ja active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007280516A (ja) | 2006-04-06 | 2007-10-25 | Sony Corp | 記録装置、記録方法および記録プログラム、記録再生装置、記録再生方法および記録再生プログラム、編集装置、編集方法および編集プログラム、ならびに、表示方法 |
JP2011502375A (ja) * | 2007-10-10 | 2011-01-20 | 韓國電子通信研究院 | ステレオスコピックデータの保存および再生のためのメタデータ構造ならびにこれを利用するステレオスコピックコンテンツファイルの保存方法 |
JP2010157826A (ja) * | 2008-12-26 | 2010-07-15 | Victor Co Of Japan Ltd | 画像復号装置、画像符復号方法およびそのプログラム |
WO2010085361A2 (en) * | 2009-01-26 | 2010-07-29 | Thomson Licensing | Frame packing for video coding |
WO2011024373A1 (ja) * | 2009-08-31 | 2011-03-03 | パナソニック株式会社 | 立体視制御装置、集積回路、立体視制御方法 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2688303A4 |
Also Published As
Publication number | Publication date |
---|---|
TW201304540A (zh) | 2013-01-16 |
JP2012249137A (ja) | 2012-12-13 |
KR20140030202A (ko) | 2014-03-11 |
EP2688303A1 (en) | 2014-01-22 |
US20140072271A1 (en) | 2014-03-13 |
CN103548345A (zh) | 2014-01-29 |
EP2688303A4 (en) | 2014-11-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5231563B2 (ja) | 立体映像データを記録する方法 | |
US8259162B2 (en) | Method and apparatus for generating stereoscopic image data stream for temporally partial three-dimensional (3D) data, and method and apparatus for displaying temporally partial 3D data of stereoscopic image | |
US8780173B2 (en) | Method and apparatus for reducing fatigue resulting from viewing three-dimensional image display, and method and apparatus for generating data stream of low visual fatigue three-dimensional image | |
US9191645B2 (en) | Stereoscopic video and audio recording method, stereoscopic video and audio reproducing method, stereoscopic video and audio recording apparatus, stereoscopic video and audio reproducing apparatus, and stereoscopic video and audio recording medium | |
RU2552137C2 (ru) | Точки входа для ускоренного 3d-воспроизведения | |
JP2011223482A (ja) | 画像処理装置、画像処理方法、およびプログラム | |
JP2006191357A (ja) | 再生装置および再生プログラム | |
KR20100068360A (ko) | 스테레오스코픽 영상 비트스트림 생성 방법 및 장치 | |
US9118895B2 (en) | Data structure, image processing apparatus, image processing method, and program | |
US10037335B1 (en) | Detection of 3-D videos | |
US20140354770A1 (en) | Digital broadcast receiving method for displaying three-dimensional image, and receiving device thereof | |
US9549167B2 (en) | Data structure, image processing apparatus and method, and program | |
WO2012165218A1 (ja) | 記録装置、記録方法、再生装置、再生方法、プログラム、および記録再生装置 | |
WO2013146636A1 (ja) | 画像符号化装置、画像復号装置、画像符号化方法、画像復号方法およびプログラム | |
JP2013051637A (ja) | 符号化装置及び符号化方法 | |
WO2012161077A1 (ja) | 記録装置、記録方法、再生装置、再生方法、プログラム、および記録再生装置 | |
KR101382618B1 (ko) | 콘텐츠 정보 생성 방법 및 콘텐츠 정보를 이용한 콘텐츠처리 장치 | |
WO2012165160A1 (ja) | 再生装置、再生方法、およびプログラム | |
EP2837183A2 (en) | Depth signaling data | |
JP2013211777A (ja) | 画像符号化装置、画像復号装置、画像符号化方法、画像復号方法およびプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12793581 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2012793581 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14118081 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 20137030803 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |