WO2014025239A1 - 3d 영상을 위한 영상 컴포넌트 송수신 처리 방법 및 장치 - Google Patents
3d 영상을 위한 영상 컴포넌트 송수신 처리 방법 및 장치 Download PDFInfo
- Publication number
- WO2014025239A1 WO2014025239A1 PCT/KR2013/007226 KR2013007226W WO2014025239A1 WO 2014025239 A1 WO2014025239 A1 WO 2014025239A1 KR 2013007226 W KR2013007226 W KR 2013007226W WO 2014025239 A1 WO2014025239 A1 WO 2014025239A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- component
- image component
- video
- information
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/44—Decoders specially adapted therefor, e.g. video decoders which are asymmetric with respect to the encoder
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/005—Aspects relating to the "3D+depth" image format
Definitions
- the present invention relates to a method and apparatus for providing a 3D broadcasting service and receiving and processing the same. More particularly, the present invention relates to a method and apparatus for decoding a texture image component and a depth map image component required for 3D image rendering.
- 3DTV three-dimensional television
- a 3D image provides a stereoscopic effect using the stereo vision principle of two eyes. Humans feel perspective through the parallax of two eyes, that is, binocular parallax by the distance between two eyes about 65mm apart, so the 3D image provides an image so that each of the left and right eyes can see the associated plane image. It can provide three-dimensional and perspective.
- Such 3D image display methods include a stereoscopic method, a volumetric method, a holographic method, and the like.
- a left view (left view) image for viewing in the left eye and a right view (light view) image for viewing in the right eye are provided, and the left eye through polarizing glasses or the display device itself is provided.
- the right eye and the right eye respectively recognize the 3D image effect by viewing the left view image and the right image.
- the viewer's position may change every moment.
- the current 3DTV receiver can feel a proper 3D effect only when watching the 3D content from a fixed viewpoint or position, and thus, the viewer may have a problem of viewing the 3D content at a limited position.
- the technical problem to be solved by the present invention is to solve the above-described problem, and to allow a viewer to freely select a desired view point and consume 3D content while feeling an appropriate 3D effect. More specifically, the technical task of the present invention is to provide an appropriate broadcast signaling structure, so that the receiver can generate virtual 3D images according to each viewpoint that can vary depending on the viewer.
- a method of processing a component of a video component for 3D video includes generating at least one texture image component and at least one depth map image component, between the at least one texture image component and at least one depth map image component.
- SEI Supplemental Enhancement Information
- the 3D layer information element includes layer image type information for identifying whether a specific image component corresponds to a texture image component or a depth map image component.
- the 3D layer information element further includes layer image identifier information indicating a viewpoint of the specific image component.
- the 3D layer information element further includes embedded layer image number information indicating the number of image components to be decoded before decoding the specific image component.
- the 3D layer information element includes embedded layer image type information for identifying whether an image component to be decoded corresponds to a texture image component or a depth map image component before decoding the specific image component, and The method further includes coding priority information indicating a decoding priority of an image component to be decoded before decoding the specific image component.
- the method for processing video component transmission for 3D video includes generating an SEI message including a GVR information element signaling information relating to video component reconstruction using an arbitrary access point, and including the GVR information element. Generating a NAL unit including the SEI message.
- the GVR information element may include GVR identification information indicating whether a specific access unit is a GVR random access point, and in video components to which the GVR is applied, in time than the video component of the current view.
- GVR identification information indicating whether a specific access unit is a GVR random access point, and in video components to which the GVR is applied, in time than the video component of the current view.
- 3D extended video number information indicating the number of video components that cannot be decoded when there is no video component, and video identification information identifying the non-decoded video components.
- the method for transmitting a video component for a 3D video includes generating a SEI message including a 3D video dependency information element for signaling an association between the texture video component and the depth map video component, and the 3D video.
- the 3D image dependency information element includes information indicating whether an SEI message including the 3D image dependency information element provides information on all image components included in 3D content.
- an image component receiving apparatus for a 3D image comprises: a tuner for receiving a broadcast signal including at least one texture image component, the at least one depth map image component, and the NAL unit, the broadcast signal A demultiplexer for parsing a NAL unit including a Supplemental Enhancement Information (SEI) message including a 3D layer information element signaling coding related layer information between the one or more texture image components and the one or more depth map image components, and And an image component decoder for decoding the at least one texture image component and the at least one depth map image component using a 3D layer information element included in an SEI message.
- SEI Supplemental Enhancement Information
- the 3D layer information element includes layer image type information for identifying whether a specific image component corresponds to a texture image component or a depth map image component.
- the 3D layer information element further includes embedded layer image number information indicating the number of image components to be decoded before decoding the specific image component.
- the 3D layer information element includes embedded layer image type information for identifying whether an image component to be decoded corresponds to a texture image component or a depth map image component before decoding the specific image component, and The method further includes coding priority information indicating a decoding priority of an image component to be decoded before decoding the specific image component.
- the demultiplexer further parses a NAL unit including an SEI message including a GVR information element signaling information relating to video component reconstruction using an arbitrary access point
- the video component decoder is further configured to include the GVR information.
- the GVR information element indicates whether a particular access unit is a GVR arbitrary access point.
- GVR identification information, 3D extended video number information indicating the number of video components that cannot be decoded when there is no video component temporally ahead of the video component at the current time in the video components to which the GVR is applied, and the non-decoded video components Image to identify It includes specific information.
- the demultiplexer further parses a NAL unit comprising an SEI message comprising a 3D image dependent information element signaling an association between the texture image component and the depth map image component, wherein the image component decoder And further using the 3D image dependent information element to decode the at least one texture image component and the at least one depth map image component, wherein the 3D image dependent information element comprises the 3D image dependent information element. It includes information indicating whether the message provides information on all image components included in the 3D content.
- an image component transmission processing method for a 3D image may include generating at least one texture image component and at least one depth map image component, between the at least one texture image component and at least one depth map image component.
- embedded layer image number information indicating the number of image components to be decoded before decoding the specific image component, wherein the one or more texture image components, the one or more depth map image components, and an image are displayed.
- a step of generating a broadcast signal including 3D hierarchy descriptor including 3D hierarchy descriptor.
- the 3D layer information descriptor is included in PMT, VCT, or SDT.
- the 3D layer information descriptor may include embedded layer image type information for identifying whether an image component to be decoded corresponds to a texture image component or a depth map image component before decoding the specific image component.
- the method further includes coding priority information indicating a decoding priority of an image component to be decoded before decoding a specific image component.
- a new left image and a right image can be generated in the receiver, and the stereoscopic effect can be adjusted using this combination.
- FIG. 1 is a diagram illustrating a part of a receiver for generating a 3D image of an arbitrary view according to an embodiment of the present invention.
- FIG 2 illustrates video data and depth elements for 3D content, in accordance with an embodiment of the invention.
- FIG. 3 is a diagram illustrating a coding mode for coding a 3D image according to an embodiment of the present invention.
- FIG. 4 is a diagram illustrating a structure of an encoder according to a coding mode according to an embodiment of the present invention.
- FIG. 5 is a diagram illustrating a procedure of coding a texture image and a depth map image according to an embodiment of the present invention.
- FIG. 6 is a diagram illustrating a coding order of each image component in one access unit according to an embodiment of the present invention.
- FIG. 7 illustrates a gradual view refresh (GVR) access unit according to an embodiment of the present invention.
- GVR gradual view refresh
- FIG. 8 is a diagram illustrating decoding using stepped screen switching (GVR) according to an embodiment of the present invention.
- FIG 9 illustrates a 3D view dependency descriptor according to an embodiment of the present invention.
- FIG. 10 is a diagram illustrating a value of a depth_hor_upsampling_factor field and a depth_ver_upsampling_factor field according to an embodiment of the present invention.
- FIG. 11 is a diagram illustrating a location of a 3D view dependency descriptor in a PMT according to an embodiment of the present invention.
- FIG. 12 illustrates a 3D hierarchy descriptor according to an embodiment of the present invention.
- FIG. 13 is a diagram illustrating values of a hor_upsampling_factor field and a ver_upsampling_factor field according to an embodiment of the present invention.
- FIG. 14 illustrates a location of a 3D hierarchy descriptor in a PMT according to an embodiment of the present invention.
- 15 is a diagram illustrating meaning according to a value of a stream_type field according to an embodiment of the present invention.
- 16 is a diagram illustrating signaling of a video level, according to an embodiment of the present invention.
- 17 is a diagram illustrating syntax of an SEI message when 3D view dependency info is transmitted through an SEI payload according to an embodiment of the present invention.
- FIG. 18 is a diagram illustrating syntax of an SEI message when a 3D hierarchy info is transmitted through an SEI payload according to an embodiment of the present invention.
- 19 is a diagram illustrating syntax of an SEI message when 3D hierarchy info is transmitted through an SEI payload according to an embodiment of the present invention.
- FIG. 20 is a diagram illustrating the contents of a NAL unit and a syntax structure of an RBSP according to a value of nal_unit_type according to an embodiment of the present invention.
- 21 is a diagram for describing a decoding operation of a receiver using 3D hierarchy Info according to an embodiment of the present invention.
- FIG. 22 illustrates a 3D hierarchy Info for decoding operation of the receiver of FIG. 21 according to an embodiment of the present invention.
- FIG. 23 is a diagram showing the structure of a 3DTV receiver according to an embodiment of the present invention.
- FIG. 1 is a diagram illustrating a part of a receiver for generating a 3D image of an arbitrary view according to an embodiment of the present invention.
- a left image (or a left video) of stereo 3D content a texture view constituting a right image (or a right video), and a depth map corresponding to each texture image
- a depth map view a texture view constituting a right image (or a right video)
- the texture image corresponds to an image image composed of pixels.
- Texture image 0 may be used to generate a left image necessary for reproducing 3D content
- texture view 1 may be used to generate a right image required for reproducing 3D content.
- the depth map image is an image or data including information for giving a depth to each of the left image and the right image constituting the 3D content.
- the depth map image may include depth information of specific points in each image of the left image and the right image constituting the 3D content.
- the depth map image may include depth information about each pixel in the image, and may include depth information about pixels of a specific unit.
- the depth map image may provide depth information about an object included in the image.
- the depth map image may include depth information about one or more viewpoints. That is, the depth map image may include one or more depth information about a pixel, a set of pixels, or objects in the image, and each depth information may be depth information corresponding to each viewpoint.
- the image synthesis unit may generate a left image and a right image included in the 3D content by using the texture image (the texture image for the left image and the texture image for the right image) and the depth map image.
- the image analyzer may use a camera parameter including environment information of a camera photographing the 3D content in addition to the depth map image.
- the camera parameter may include a distance between the camera, a distance between the camera and a photographing target, an angle between the cameras, a height of the camera, and / or setting information of the camera itself.
- the 3D formatter generates a 3D image by rendering the left image and the right image generated by the image analyzer.
- FIG 2 illustrates video data and depth elements for 3D content, in accordance with an embodiment of the invention.
- depth information about the texture image and the texture image may be provided.
- depth information about the texture image and the texture image may be provided in one program unit, service unit, or channel unit.
- the depth information may be the above-described depth map image.
- the receiver needs to distinguish between the depth information of the texture image and the texture image, and determine which depth information is applied to which texture image. Therefore, it is necessary to transmit signaling information about this through a broadcast signal.
- FIG. 3 is a diagram illustrating a coding mode for coding a 3D image according to an embodiment of the present invention.
- the 3D video may be coded with AVC (Advanced Video Coding).
- 3D video may be coded by Multiview Video Coding (MVC).
- MVC Multiview Video Coding
- the 3D image may be coded in at least two modes.
- the “3D High” mode is a mode in which AVC and / or MVC, which was applied only to a texture image (or texture data), is also applied to depth data (or depth map image and depth information).
- the texture image can be compatible with the MVC codec.
- the base layer image may be compatible with both AVC and MVC codec. According to this mode, coding of the texture image and coding of the depth data are independently performed.
- the “3D Extended High” mode is a method in which the coding of the texture image and the depth data based on AVC and / or MVC is not independent but mutually influenced and coded.
- the texture image can be compatible with the AVC codec.
- the base layer image may be compatible with both AVC and MVC codec. This mode is generally better coding efficiency than the "3D High” mode.
- FIG. 4 is a diagram illustrating a structure of an encoder according to a coding mode according to an embodiment of the present invention.
- FIG. 4A shows the structure of an encoder according to the “3D High” mode. It may be coded with AVC and / or MVC of the texture pictures T0 and T1.
- the depth data D0 and D1 may be coded in AVC and / or MVC separately from the texture image, and there is no dependency between the depth data D0 and D1 and the texture image T0 and T1, thereby providing depth data.
- Each of the (D0, D1) and the texture image (T0, T1) may be separately decoded.
- the texture images T0 and T1 and the depth data D0 and D1 may be coded with AVC and / or MVC. However, coding of the texture image and coding of the depth data are performed with dependencies on each other. Therefore, the texture image and the depth data must be decoded together.
- the base texture view (T0) may be compatible with the AVC codec in any mode.
- FIG. 5 is a diagram illustrating a procedure of coding a texture image and a depth map image according to an embodiment of the present invention.
- the bit stream may be encoded for each video component, and the decoding order may be performed for each video component. That is, the texture image component for the left image (texture of view 0), the component for the right image (texture of view 1), the depth map image for the left image (depth map of view 0), and the depth map for the right image An image (depth map of view 1) may be encoded and transmitted in each bit stream, and processing of data of another image component may not be performed before processing of data of one image component is terminated.
- the bit stream may be encoded for each access unit, and the decoding order may be performed for each access unit. That is, one left image frame, one right image frame, one depth map frame for one left image frame, and one depth map frame for one right image frame may be included in one access unit.
- the unit may be encoded and transmitted in a bit stream. Before the processing for data included in one access unit is finished, the processing for data included in another access unit may not be performed.
- the texture image component of the AVC / MVC compatible image may be coded to be located before the corresponding depth map image component. That is, the texture image 0 may be coded to be positioned first among image components for a specific viewpoint.
- the encoding may be encoded according to the dependency of the texture image component and the depth map image component.
- FIG. 6 is a diagram illustrating a coding order of each image component in one access unit according to an embodiment of the present invention.
- texture image 0 is encoded first, and texture image 1 having a dependency thereon is encoded. Thereafter, depth map image 0 for texture image 0 is encoded, and depth map image 1 for texture image 1 with dependency thereon is encoded.
- both texture image 0 and texture image 1 may be texture images compatible with AVC / MVC.
- texture image 0 is encoded first, and depth map image 0 for texture image 0 is encoded. Thereafter, texture image 1 is encoded with dependency on texture image 0, and depth map image 1 for texture image 1 is encoded with dependency with depth map image 0.
- both texture image 0 and texture image 1 may be texture images compatible with AVC / MVC.
- texture image 0 is encoded first, and depth map image 0 for texture image 0 is encoded.
- the depth map image 0 may have a dependency on the texture image 0.
- the depth map image 1 for the texture image 1 is encoded.
- the depth map image 1 may have a dependency on the depth map image 0.
- texture image 1 is encoded.
- Texture image 1 may have a dependency on texture image 0.
- the texture image 1 may have a dependency on the depth map image 1.
- texture image 0 may be AVC / MVC compatible, but texture image 1 may not be compatible with MVC.
- the fact that the B picture has a dependency on the A picture means that a specific element of the A picture may be used in the encoding process of the B picture. In this case, the A picture must be decoded first in the decoding process so that the B picture can be completely decoded.
- encoding of the B picture may be performed by encoding data about a difference between the original A picture and the original B picture.
- encoding of the B picture may be performed by encoding additional information about some pixels of the A picture.
- FIG. 7 illustrates a gradual view refresh (GVR) access unit according to an embodiment of the present invention.
- GVR gradual view refresh
- the GVR precisely decodes the pre-selected viewpoints and allows the rest of the viewpoints to be recovered using a synthesis technique.
- GVR is a kind of random access point that can be used in 3D video coding and has a high compression efficiency compared to the existing Random access point Instantaneous Decoding Refresh (IDR).
- the random access point refers to a function for reproducing a video (or a frame) at a corresponding time point or after the corresponding time point when data is received in the middle of a bit string of the video data.
- the left and right images constituting the 3D image may be immediately acquired at the corresponding time point.
- the GVR may not immediately acquire the left and right images at that point in time, and may include one of the left and right images, or Part of the left or right image can be acquired. Thereafter, both left and right images may be acquired gradually. That is, according to the GVR, a complete left image and a right image may be acquired from a specific time point after the receiver approaches the bit string.
- inter prediction is not used in IDR and anchor access units, some limited inter prediction is used for GVR access units, and image synthesis is used.
- signaling of the GVR access unit may be transmitted through an SEI message and a NAL unit header.
- Images belonging to each count may be reconstructed through prediction using images belonging to adjacent counts.
- a P frame may be derived from an I frame (I picture) in one GVR access unit.
- each P frame for a left image and a right image may be obtained using an I frame included in one GVR access unit.
- FIG. 8 is a diagram illustrating decoding using stepped screen switching (GVR) according to an embodiment of the present invention.
- a texture image component and a depth map image component to which inter prediction is not applied are first decoded.
- “Non-refreshed” images are reconstructed through image analysis (eg depth-image-based rendering; DIBR).
- DIBR depth-image-based rendering
- the image reconstructed through the image analysis may have an approximate shape without completely reconstructing the original image.
- the complete image may be restored. For example, in FIG. 8, the complete image is restored when the images corresponding to the count 30 are reached.
- FIG 9 illustrates a 3D view dependency descriptor according to an embodiment of the present invention.
- the 3D view dependency descriptor according to the present invention may be included in a program map table (PMT), a virtual channel table (VCT), or an event information table (EIT) of an ATSC system.
- PMT program map table
- VCT virtual channel table
- EIT event information table
- the 3D view dependency descriptor may be included in a service description table (SDT) or an event information table (EIT) of the DVB system and transmitted. According to the table in which the 3D view dependency descriptor is transmitted, it is determined whether signaling is performed at the service level, the program level, or the event level.
- the 3D view dependency descriptor signals a connection relationship between the texture image and the depth map image included in the broadcast program, the broadcast channel, or the broadcast service, and provides information about dependency in coding between the images.
- the 3D view dependency descriptor may include information on whether the GVR technique is used as an arbitrary access point.
- the receiver When reconstructing an image using GVR, the receiver must support depth-image-based rendering (DIBR) function, and thus can provide information on whether the GVR technique is used at the system level. In this case, corresponding information may be transmitted through an SEI message or a NAL unit header of video data.
- DIBR depth-image-based rendering
- the 3D view dependency descriptor may include a 3D_mode field, view_ID field, view_PID field, independent_texture_coding_flag field, existence_of_corresponding_depth field, depth_PID field, depth_hor_upsampling_factor field, depth_ver_upsampling_factor field, GVR_included field, and / or cam_param_included.
- the 3D_mode field is a field indicating coding dependency information between the texture image and the depth map image. For example, if this field is 0, it indicates the “3D High” mode (if the texture image and the depth map image are coded independently of each other). If this field is 1, it indicates the “3D Extended High” mode (the texture image and the depth map. High coding efficiency due to dependence between images).
- a receiver that wants to acquire only the texture image can decode the texture image without decoding the depth map image.
- the receiver should utilize the decoding result of the depth map image when decoding the texture image. Or, the receiver should utilize the decoding result of the texture image when decoding the depth map image.
- the view_ID field is an identifier indicating a viewpoint of the texture video component.
- the view_ID value for the depth map image component has the same value as that of the view_ID value of the texture image component.
- the view_PID field is a PID value of an ES stream for transmitting a texture video component.
- the independent_texture_coding_flag field represents whether independent coding for each texture image is possible. For example, since texture image 0 (T0) can be independently coded by default, this field is not necessarily required, and when used, it has a value of '1'. For the remaining texture images (eg, T1), it may be determined whether independent decoding is possible through this field.
- the existence_of_corresponding_depth field is a field indicating whether a depth map image component corresponding to the texture image component is signaled. For example, a value of '0' indicates that only a texture image component exists, and a value of '1' indicates that both a texture image component and a depth map image component exist.
- the depth_PID field is a value signaled when the existence_of_corresponding_depth field is '1' and is a PID value of an ES stream for transmitting a depth map image component corresponding to a texture image component.
- the depth_hor_upsampling_factor field is a field for indicating information of a horizontal upsampling factor required in a process of restoring an image after the depth map image component corresponding to the texture image component is decoded.
- the depth_ver_upsampling_factor field is a field for indicating information of a vertical upsampling factor required in a process of restoring an image after the depth map image component corresponding to the texture image component is decoded.
- the transmitter may rescale and transmit the resolution of the depth map image component.
- the receiving end needs to be rescaled when applying the corresponding depth map image in 3D image analysis.
- the receiver may process the reference by referring to an upsampling factor (eg, depth_hor_upsampling_factor field or depth_ver_upsampling_factor field). This upsampling factor can also be specified for texture image components, if necessary.
- depth_hor_upsampling_factor field and / or the depth_ver_upsampling_factor field may have and a description thereof will be described later.
- the GVR_included field represents whether an image can be restored by the GVR technique. For example, if the GVR_included field is 0, it indicates that the GVR picture is included in the program, channel, or service. If the GVR_included field is 1, the GVR picture is not included in the program, channel, or service. Indicates. Since the GVR can be used as a random access point only when DIBR function is provided to the receiving end, receivers that do not support DIBR use this field value to enable random access point of the current program, channel or service. It is possible to determine in advance whether all points) are accessible or not.
- the cam_param_included field is a field indicating whether a camera parameter is signaled for a specific image. Camera parameters may be required for the GVR to be used as a random access point. In this case, the receiver may know whether GVR support is possible through this field. In general, when a broadcaster sends a GVR, camera parameters can also be sent, but if the camera parameters are sent by a separate path other than the video header, the receiver can use this field to identify whether or not the camera parameters are provided. Can be.
- All texture images may be signaled using a for loop of the texture image included in the 3D View dependency descriptor, or only some texture images may be signaled.
- the number of for loops may be executed by including the num_of_views field in the 3D View dependency descriptor.
- the 3D_mode field is 1, that is, in the “3D Extended High” mode, a method of informing hierarchical information between image components in terms of decoding order is required. That is, there must be a signaling method for notifying at the system level the dependency information between the texture image component and the depth map image component included in the program, channel, or service, which is to extend Hierarchy_descriptor, which is a descriptor of the existing MPEG-2 system, Alternatively, a method of signaling by including related information in a 3D view dependency descriptor or a method of separately designating and signaling a 3D hierarchy descriptor may be used. More details will be described later.
- FIG. 10 is a diagram illustrating a value of a depth_hor_upsampling_factor field and a depth_ver_upsampling_factor field according to an embodiment of the present invention.
- a value of the depth_hor_upsampling_factor field indicates that the upsampling factor is applied in the horizontal direction of the image frame, and a value of the depth_ver_upsampling_factor field indicates that the corresponding upsampling factor is applied in the vertical direction of the image frame.
- the decoder may separately check information included in the video elementary stream to obtain accurate upsampling factor information.
- the value of the depth_hor_upsampling_factor field or the depth_ver_upsampling_factor field is '0011', it indicates that the coding resolution of the depth map image component is 3/4 of the coding resolution of the texture image component.
- FIG. 11 is a diagram illustrating a location of a 3D view dependency descriptor in a PMT according to an embodiment of the present invention.
- the PMT includes the following.
- It may include a table_id field, section_syntax_indicator field, section_length field, program_number field, version_number field, current_next_indicator field, section_number field, last_section_number field, PCR_PID field, program_info_length field, stream_type field, elementary_PID field, ES_info_length field and / or CRC_32 field.
- the table_id field is an 8-bit field and may be set to a value of 0x02 to indicate that it is TS_program_map_section.
- the section_syntax_indicator field is set to 1 as 1 bit.
- the section_length field consists of 12 bits and the first two bits are 00. This field indicates the number of bytes in the section, and indicates the length from this field to the CRC.
- the program_number field consists of 16 bits. This indicates to which program the program_map_PID is applicable.
- One program definition is transmitted by only one TS_program_map_section.
- the version_number field represents the version of the virtual channel. Increment by 1 for every change in the VCT. When the version value reaches 31, the next version value is 0. The value of this field may have the same value as the same field value of the MGT.
- the current_next_indicator field consists of 1 bit and the value is set to 1 when the VCT is currently applicable. If it is set to 0, this means that it is not applicable yet and the following table is valid.
- the section_number field identifies the number of the section. For example, the value of this field for PMT is set to 0x00.
- the last_section_number field identifies the number of the last section. For example, the value of this field for PMT is set to 0x00.
- the PCR_PID field means a PID of a TS that consists of 13 bits and includes a PCR field valid for a program described by a program number.
- the program_info_length field may consist of 12 bits, and the first two bits have a value of 00. The remaining 10 bits represent the descriptor following this field in number of bytes.
- the stream_type field may consist of 8 bits and indicates the type of a program element transmitted by a packet having a PID value of a basic PID.
- the elementary_PID field may consist of 13 bits and indicate a PID of a TS including a related program element.
- the ES_info_length field may consist of 12 bits, and the first two bits are 00. The remaining 10 bits represent the descriptor of the associated program element following this field in number of bytes.
- the descriptor following the ES_info_length field may be a 3DTV service location descriptor.
- the CRC_32 field indicates a CRC value that causes a zero output of a register in the decoder.
- the 3D view dependency descriptor may be included in an area where a program level descriptor may be located in the PMT. That is, the 3D view dependency descriptor may be included in the for loop below the program_info_length field.
- the 3D View dependency descriptor when transmitted to the VCT, it may be located in a channel level descriptor loop of the VCT.
- the 3D View dependency descriptor when the 3D View dependency descriptor is transmitted to the SDT, the 3D View dependency descriptor may be located in the service level descriptor loop and may be set to include information on all images in the 3D View dependency descriptor.
- FIG. 12 illustrates a 3D hierarchy descriptor according to an embodiment of the present invention.
- the 3D hierarchy descriptor includes coding related hierarchical information of a texture image component and a depth map image component constituting a program in a broadcasting system based on 3D video coding.
- the 3D hierarchy descriptor may be located at an elementary stream level of the PMT.
- the 3D hierarchy descriptor may include information of an image component required for decoding the corresponding ES stream.
- the 3D hierarchy descriptor may include a hierarchy_view_type field, a hierarchy_view_ID field, a num_hierarchy_embeded_views field, a hierarchy_embeded_view_type field, a hierarchy_embeded_view_ID field, a coding_priority_index field, a hor_upsampling_factor field, and / or a ver_upsampling_factor field.
- the hierarchy_view_type field is a field indicating the type of a video component. For example, if the value of this field is '0', it means that the image component is a texture image component. If the value of this field is '1', it means that the image component is a depth map image component.
- the hierarchy_view_ID field is a view identifier indicating a viewpoint of an image component.
- the num_hierarchy_embeded_views field is a field indicating the number of video components to be decoded before decoding the video components. For example, when there is a dependency between video components, there may be a video component necessary to decode the current video component. In this case, the receiver may know the number of video components required through the num_hierarchy_embeded_views field.
- the hierarchy_embeded_view_type field is a field for indicating what kind of video component should be decoded before decoding the video component. For example, if the value of the hierarchy_embeded_view_type field is '0', it means a texture image component. If the value of the hierarchy_embeded_view_type field is '1', it means 'depth map image.'
- the hierarchy_embeded_view_ID field is a view identifier indicating a viewpoint of an image component to be decoded before decoding the image component.
- the receiver may find an image component necessary for decoding the current image component through the hierarchy_embeded_view_ID field. Alternatively, the receiver may find the necessary image component by using a combination of the hierarchy_embeded_view_type field and the hierarchy_embeded_view_ID field.
- the coding_priority_index field is a field indicating a decoding priority of a video component to be decoded before decoding the video component. This field may be replaced with the value of the view_index field. This field is a field for signaling the priority when there are two or more streams to be decoded first.
- the hor_upsampling_factor field is a field for indicating information of a horizontal upsampling factor required in a process of restoring an image component after the image component is decoded.
- the ver_upsampling_factor field is a field for indicating information of a vertical upsampling factor required in a process of restoring an image component after the image component is decoded.
- the transmitter may rescale and transmit the resolution of an image component.
- the receiving end needs to be rescaled when the corresponding video component is applied, and can be processed by referring to an upsampling factor (eg, a hor_upsampling_factor field and / or a ver_upsampling_factor field).
- an upsampling factor eg, a hor_upsampling_factor field and / or a ver_upsampling_factor field.
- FIG. 13 is a diagram illustrating values of a hor_upsampling_factor field and a ver_upsampling_factor field according to an embodiment of the present invention.
- a value of the hor_upsampling_factor field indicates that the upsampling factor is applied in the horizontal direction of the image frame
- a value of the ver_upsampling_factor field indicates that the upsampling factor is applied in the vertical direction of the image frame.
- the decoder may separately check information included in the video elementary stream to obtain accurate upsampling factor information.
- the hor_upsampling_factor field or the ver_upsampling_factor field When the value of the hor_upsampling_factor field or the ver_upsampling_factor field is '0011', it indicates that the coding resolution of the image component is 3/4 of the coding resolution of the base texture image component.
- FIG. 14 illustrates a location of a 3D hierarchy descriptor in a PMT according to an embodiment of the present invention.
- the 3D hierarchy descriptor may be included in an area where a program lever descriptor of the PMT is located.
- an additional stream_type field may also be newly signaled. That is, since streams related to 3D coding are new streams that were not present in the existing AVC / MVC, the new stream_type field is assigned a value, and receivers compatible only with the existing AVC / MVC are designed to ignore the stream corresponding to the value of the specific stream_type field. As a result, backward compatibility between the new signaling system and the existing system can be maintained.
- 15 is a diagram illustrating meaning according to a value of a stream_type field according to an embodiment of the present invention.
- 3D enhancement streams can be classified into three categories.
- the AVC / MVC compatible depth map image component may be divided into an AVC compatible depth map image component and an MVC compatible depth map image component.
- the value of the stream_type field may be set to '0x24'.
- the value of the stream_type field may be set to '0x25'.
- the value of the stream_type field may be set to '0x26'.
- the stream_type field may indicate that the texture video component is coded in a new manner, which is not compatible with AVC or MVC.
- the texture image component may be coded by inter-layer prediction using a depth map image component.
- the value of the stream_type field may be set to '0x27'.
- the stream_type field may indicate that the depth map image component is coded in a new manner that is not AVC or MVC compatible.
- the depth map image component may be coded by inter-layer prediction using the texture image component.
- a 3D Hierarchy descriptor may be transmitted to the VCT. Basically, it can be designed to have the same function as the 3D Hierarchy descriptor defined in PMT, and the following can be additionally considered.
- the 3D Hierarchy descriptor When the 3D Hierarchy descriptor is included in the VCT, since it is a channel level descriptor, the level described is different from the 3D Hierarchy descriptor of the PMT signaled for each element stream. That is, the 3D Hierarchy descriptor included in the PMT needs a different configuration from the ES level descriptor. Therefore, in order to signal the characteristics of each stream included in the channel, a method such as listing the characteristics of each component by including a for loop in the 3D Hierarchy descriptor can be used. In this case, a field commonly applied to all streams is for You can pull it out of the loop.
- the receiver may determine which ES stream the information included in the 3D Hierarchy descriptor is applied to through the added field.
- the 3D Hierarchy descriptor may be transmitted in the SDT. In this case, it can basically be designed to have the same function as the 3D Hierarchy descriptor defined in the PMT, and the following may be additionally considered.
- the 3D Hierarchy descriptor is a service level descriptor
- the level described in the 3D Hierarchy descriptor of the PMT signaled for each element stream is different. That is, the 3D Hierarchy descriptor included in the PMT needs a different configuration from the ES level descriptor. Therefore, in order to signal the characteristics of each stream included in the channel, a method such as listing the characteristics of each component by including a for loop in the 3D Hierarchy descriptor can be used. In this case, a field commonly applied to all streams is You can pull it out of the loop.
- a field including a component_tag value of the corresponding stream may be added. Through this field, the receiver can determine to which ES stream the information included in the 3D Hierarchy descriptor is applied.
- the content corresponding to the newly designated stream_type field may be newly designated and may be signaled to be represented by a combination of the stream_content field and the component_type field of the component descriptor.
- 16 is a diagram illustrating signaling of a video level, according to an embodiment of the present invention.
- information included in the above-described 3D View dependency descriptor, information included in the 3D Hierarchy descriptor, and / or GVR related information may be transmitted at the video level.
- the information may be transmitted to a Supplemental Enhancement Information (SEI) area including information to assist processes such as decoding and output.
- SEI Supplemental Enhancement Information
- 3D_view_dependency_info SEI message, 3D_hierarchy_info SEI message and GVR_info SEI message which are entered into the video level for transmission of 3D View dependency Info, 3D Hierarchy Info, and GVR Info information in SEI, respectively, or 3D View dependency Info, 3D Hierarchy Info, It is also possible to specify two or more pieces of GVR Info information by integrating them into one SEI message.
- the SEI payloadType element When transmitting 3D View dependency Info, 3D Hierarchy Info, and GVR Info in each SEI message, it is possible to signal what information the SEI payload includes through the value of the payloadType element. For example, when the value of the payloadType element is '48', it may represent that the 3D view dependency info is transmitted through the SEI payload. When the value of the payloadType element is '49', it may represent that the 3D hierarchy info is transmitted through the SEI payload. When the value of the payloadType element is '50', it may represent that GVR info is transmitted through the SEI payload. In addition, the payloadSize element may signal the size of information included in the SEI payload.
- 17 is a diagram illustrating syntax of an SEI message when 3D view dependency info is transmitted through an SEI payload according to an embodiment of the present invention.
- the SEI message for transmitting the 3D view dependency info includes a 3D_view_dependency_info_id field, a 3D_complete_views_or_not_flag field, a 3D_mode field, a view_ID field, an independent_texture_coding_flag field, an existence_of_corresponding_depth field, a GVR_included field, and / or include / includedparameter, and / or include / includedparameter.
- the 3D_view_dependency_info_id field represents an identification number for identifying a combination of information included in an SEI message for transmitting 3D view dependency info.
- the 3D_complete_views_or_not_flag field is a field indicating whether the content included in the SEI message is applied to all images included in the current program (or channel, service, or operation point).
- the 3D_mode field is a field indicating coding dependency information between the texture image and the depth map image. For example, if this field is 0, it indicates the “3D High” mode (if the texture image and the depth map image are coded independently of each other). If this field is 1, it indicates the “3D Extended High” mode (the texture image and the depth map. High coding efficiency due to dependence between images).
- a receiver that wants to acquire only the texture image can decode the texture image without decoding the depth map image.
- the receiver should utilize the decoding result of the depth map image when decoding the texture image. Or, the receiver should utilize the decoding result of the texture image when decoding the depth map image.
- the view_ID field is an identifier indicating a viewpoint of the texture video component.
- the view_ID value for the depth map image component has the same value as that of the view_ID value of the texture image component.
- the independent_texture_coding_flag field represents whether independent coding for each texture image is possible. For example, since texture image 0 (T0) can be independently coded by default, this field is not necessarily required, and when used, it has a value of '1'. For the remaining texture images (eg, T1), it may be determined whether independent decoding is possible through this field.
- the existence_of_corresponding_depth field is a field indicating whether a depth map image component corresponding to the texture image component is signaled. For example, a value of '0' indicates that only a texture image component exists, and a value of '1' indicates that both a texture image component and a depth map image component exist.
- the GVR_included field represents whether an image can be restored by the GVR technique. For example, if the GVR_included field is 0, it indicates that the GVR picture is included in the program, channel, or service. If the GVR_included field is 1, the GVR picture is not included in the program, channel, or service. Indicates. Since the GVR can be used as a random access point only when DIBR function is provided to the receiving end, receivers that do not support DIBR use this field value to enable random access point of the current program, channel or service. It is possible to determine in advance whether all points) are accessible or not.
- the cam_param_included field is a field indicating whether a camera parameter is signaled for a specific image. Camera parameters may be required for the GVR to be used as a random access point. In this case, the receiver may know whether GVR support is possible through this field. In general, when a broadcaster sends a GVR, camera parameters can also be sent, but if the camera parameters are sent by a separate path other than the video header, the receiver can use this field to identify whether or not the camera parameters are provided. Can be.
- FIG. 18 is a diagram illustrating syntax of an SEI message when a 3D hierarchy info is transmitted through an SEI payload according to an embodiment of the present invention.
- the SEI message for transmitting the 3D hierarchy info may include a 3D_hierarchy_info_id field, a hierarchy_view_type field, a hierarchy_view_ID field, a num_hierarchy_embeded_views field, a hierarchy_embeded_view_type field, a hierarchy_embeded_view_ID field, and / or a coding_priority_index field.
- the 3D_hierarchy_info_id field represents an identification number for identifying a combination of information included in an SEI message for transmitting 3D hierarchy info.
- the hierarchy_view_type field is a field indicating the type of a video component. For example, if the value of this field is '0', it means that the image component is a texture image component. If the value of this field is '1', it means that the image component is a depth map image component.
- the hierarchy_view_ID field is a view identifier indicating a viewpoint of an image component.
- the num_hierarchy_embeded_views field is a field indicating the number of video components to be decoded before decoding the video components. For example, when there is a dependency between video components, there may be a video component necessary to decode the current video component. In this case, the receiver may know the number of video components required through the num_hierarchy_embeded_views field.
- the hierarchy_embeded_view_type field is a field for indicating what kind of video component should be decoded before decoding the video component. For example, if the value of the hierarchy_embeded_view_type field is '0', it means a texture image component. If the value of the hierarchy_embeded_view_type field is '1', it means 'depth map image.'
- the hierarchy_embeded_view_ID field is a view identifier indicating a viewpoint of an image component to be decoded before decoding the image component.
- the receiver may find an image component necessary for decoding the current image component through the hierarchy_embeded_view_ID field. Alternatively, the receiver may find the necessary image component by using a combination of the hierarchy_embeded_view_type field and the hierarchy_embeded_view_ID field.
- the coding_priority_index field is a field indicating a decoding priority of a video component to be decoded before decoding the video component. This field may be replaced with the value of the view_index field. This field is a field for signaling the priority when there are two or more streams to be decoded first.
- 19 is a diagram illustrating syntax of an SEI message when GVR info is transmitted through an SEI payload according to an embodiment of the present invention.
- GVR generates a video corresponding to an enhancement stream through video analysis, so that the GVR acts as a random access point.
- the GVR Info SEI message can be signaled in the form of sticking to each Access Unit, which is a GVR Random Access Point, and an image that cannot be output only by the decoding process in the corresponding Access Unit (that is, a view that can be obtained only by restoring by applying DIBR at the receiver). I can tell you what the ID is).
- the GVR Info SEI message includes a GVR_flag field, a num_3D_extended_views field, and / or a view_ID field.
- the GVR_flag field identifies whether the 3D video is a GVR picture or not. That is, when the value of the GVR_flag field is '1', it means that the corresponding Access Unit is a GVR Random Access Point.
- the num_3D_extended_views field is a field indicating the number of images that cannot be decoded when there is no previous picture in time in the GVR.
- the num_3D_extended_views field represents the number of images that can be restored by applying DIBR.
- the num_3D_extended_views field may provide information on how many more images to decode in the future to obtain a complete image.
- the view_ID field is a view identifier indicating a viewpoint of an image component.
- the view_ID field plays a role of identifying an image to which DIBR is to be applied while rotating a for loop by the number of images identified by the num_3D_extended_views field.
- the view_ID field performs a for loop as many as the number of images identified by the num_3D_extended_views field, and serves to identify an image that cannot be completely restored.
- FIG. 20 is a diagram illustrating the contents of a NAL unit and a syntax structure of an RBSP according to a value of nal_unit_type according to an embodiment of the present invention.
- a value for the NAL unit of the enhanced layer may be newly assigned to the nal_unit_type value of the NAL unit header. Accordingly, the existing receiver may not recognize the corresponding value, so that 3D View dependency Info, 3D hierarchy Info, and / or GVR Info may not be used in the existing receiver.
- the information signaled by the 3D View dependency Info SEI message, the 3D hierarchy Info SEI message, and / or the GVR Info SEI message is defined in the form of nal_unit_header_3D_extension () to be signaled through nal_unit_header. Can be.
- the case where the nal_unit_type value is 21 is newly defined.
- a 3D view dependency SEI message, a 3D hierarchy SEI message, and / or a GVR info SEI message may be included in the NAL Unit and transmitted.
- the receiver When transmitting 3D View dependency Info, 3D Hierarchy Info, and / or GVR Info in the form of an SEI message to inform 3D view dependency / 3D hierarchy / GVR related information at the video level, the receiver may operate as follows.
- the receiver may receive 3D View dependency Info, 3D Hierarchy Info, and / or GVR Info via SEI raw byte sequence payload (RBSP).
- RBSP SEI raw byte sequence payload
- the receiver parses the AVC NAL unit, recognizes SEI data when the nal_unit_type value is 6, and uses the corresponding information by reading a 3D view dependency Info SEI message having a payloadType of 48.
- the receiver parses the AVC NAL unit, recognizes SEI data when the nal_unit_type value is 6, and uses the corresponding information by reading a 3D hierarchy Info SEI message having a payloadType of 49.
- the receiver parses the AVC NAL unit, recognizes that the SAL data is nal_unit_type value of 6, and uses the corresponding information by reading a GVR Info SEI message having a payloadType of 50.
- the receiver parses the 3D_view_dependency_info SEI message to obtain 3D acquisition information and dependency information of an image.
- the receiver parses the 3D_hierarchy_info SEI message to obtain 3D hierarchy information.
- the receiver parses the GVR_info SEI message to obtain GVR information.
- the receiver may operate as follows.
- the receiver parses the NAL unit from the broadcast signal, obtains slice_layer_extension_rsbp () having a nal_unit_type value of 21, and parses 3D View dependency Info, 3D Hierarchy Info, and / or GVR Info included in slice_layer_extension_rsbp () to decode 3D video. We use for.
- 21 is a diagram for describing a decoding operation of a receiver using 3D hierarchy Info according to an embodiment of the present invention.
- Video 2 in order to decode Video 2, a total of four video components (Video 0, Depth 0, Video 1, and Depth 1) must be decoded first. Therefore, the 3D Hierarchy Info of Video 2 should include information about four video components (Video 0, Depth 0, Video 1, and Depth 1).
- FIG. 22 illustrates a 3D hierarchy Info for decoding operation of the receiver of FIG. 21 according to an embodiment of the present invention.
- Video 2 corresponds to a texture video component, it indicates that the hierarchy_view_type field is a texture video.
- the hierarchy_view_ID field has a value of '2' which is a view identifier of Video 2.
- Video 2 there are four video components to be decoded first, Video 0, Depth 0, Video 1, and Depth 1, so the num_hierarchy_embeded_views field has a value of '4'.
- the 3D hierarchy Info loops through a for loop by the value indicated by the num_hierarchy_embeded_views field, and signals information about each image component.
- the receiver may predetermine necessary texture image components and / or depth map image components when decoding a specific stream in consideration of the value of the coding_priority_index field of each image component. That is, four components V0, V1, D0, and D1 must be received to decode Video 2, and since V1 and D0 have the same coding_priority_index field value, it can be understood that parallel decoding (simultaneous decoding together) is possible.
- FIG. 23 is a diagram showing the structure of a 3DTV receiver according to an embodiment of the present invention.
- the 3DTV receiver includes a tuner & demodulator 23010, a VSB decoder 23020, a TP demultiplexer 23030, a PSI / PSIP / SI processor 23040, an image component decoder 23050, a depth map.
- the 3DTV receiver receives a broadcast signal through the tuner & demodulator 23010 and / or the VSB decoder 23020.
- the VSB decoder 23020 may be interpreted as an OFDM decoder or the like regardless of its name.
- the TP demultiplexer 23030 extracts and outputs video data, system information, and the like from the broadcast signal through filtering.
- System information may include PMT, TVCT, EIT or / and SDT.
- the TP demultiplexer 23030 may parse the SEI message and / or the NAL unit and obtain 3D View dependency Info, 3D Hierarchy Info, and / or GVR Info included in the SEI message and / or the NAL unit.
- the PSI / PSIP / SI processor 23040 may receive system information and parse the received system information to obtain 3D View dependency Info, 3D Hierarchy Info, and / or GVR Info.
- the PSI / PSIP / SI processor 23040 may parse the SEI message or the NAL unit header and obtain 3D View dependency Info, 3D Hierarchy Info, and / or GVR Info included in the SEI message or NAL unit header. That is, 3D View dependency Info, 3D Hierarchy Info, and / or GVR Info included in the SEI message or NAL unit header may be extracted from the TP demultiplexer 23030 or the PSI / PSIP / SI processor 23040.
- Image component decoder 23050 includes a depth map image component decoder 23052 and / or a texture image component decoder 23054.
- Depth map image component decoder 23052 decodes the depth map image component from the video data.
- the texture image component decoder 23054 decodes the texture image component from the video data.
- the depth map image component decoder 23052 and / or the texture image component decoder 23054 perform upsampling of image components using 3D View dependency Info, 3D Hierarchy Info, and / or GVR Info, or identify dependencies among the image components. And decode the video component according to the dependency.
- the related operation is replaced with the description of FIGS. 21 and 22.
- the image analysis module 23060 applies depth information obtained from the depth map image component to the texture image component.
- the first virtual image generator 23070 and the second virtual image generator 23080 each render a left image or a right image included in the 3D image.
- the first virtual image generator 23070 and the second virtual image generator 23080 may be included in the image analysis module 23060, and the image analysis module 23060 may render a left image or a right image according to each viewpoint. It can play a role.
- the output formatter 23090 combines a left image or a right image generated by the first virtual image generator 23070 and the second virtual image generator 23080 to output a 3D image.
- Apparatus and method according to the present invention is not limited to the configuration and method of the embodiments described as described above, the above-described embodiments may be selectively all or part of each embodiment so that various modifications can be made It may be configured in combination.
- the broadcast signal processing method related to the broadcast program of the present invention can be implemented as a processor-readable code on a processor-readable recording medium provided in the network device.
- the processor-readable recording medium includes all kinds of recording devices that store data that can be read by the processor. Examples of the processor-readable recording medium include ROM, RAM, CD-ROM, magnetic tape, floppy disk, optical data storage device, and the like, and may also be implemented in the form of a carrier wave such as transmission over the Internet. .
- the processor-readable recording medium can also be distributed over network coupled computer systems so that the processor-readable code is stored and executed in a distributed fashion.
- the present invention is applicable to a series of industries related to the provision of 3DTV broadcasting service.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
Claims (18)
- 하나 이상의 텍스쳐 영상 컴포넌트 및 하나 이상의 깊이 맵 영상 컴포넌트를 생성하는 단계;상기 하나 이상의 텍스쳐 영상 컴포넌트 및 하나 이상의 깊이 맵 영상 컴포넌트 사이의 코딩 관련 계층 정보를 시그널링하는 3D 계층 정보 엘레먼트를 포함하는 SEI 메시지를 생성하는 단계;상기 3D 계층 정보 엘레먼트를 포함하는 SEI (Supplemental Enhancement Information) 메시지를 포함하는 NAL 유닛을 생성하는 단계; 및상기 하나 이상의 텍스쳐 영상 컴포넌트, 상기 하나 이상의 깊이 맵 영상 컴포넌트, 및 상기 NAL 유닛을 포함하는 방송 신호를 생성하는 단계;를 포함하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 1 항에 있어서, 상기 3D 계층 정보 엘레먼트는,특정 영상 컴포넌트가 텍스쳐 영상 컴포넌트에 해당되는지, 깊이 맵 영상 컴포넌트에 해당되는지 여부를 식별하는 계층 영상 타입 정보를 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 2 항에 있어서, 상기 3D 계층 정보 엘레먼트는,상기 특정 영상 컴포넌트의 시점을 알려주는 계층 영상 식별자 정보를 더 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 2 항에 있어서, 상기 3D 계층 정보 엘레먼트는,상기 특정 영상 컴포넌트를 디코딩 하기 전에 디코딩되어야 하는 영상 컴포넌트의 개수를 나타내는 임베디드 계층 영상 개수 정보를 더 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 4 항에 있어서, 상기 3D 계층 정보 엘레먼트는,상기 특정 영상 컴포넌트를 디코딩 하기 전에 디코딩되어야 하는 영상 컴포넌트가 텍스쳐 영상 컴포넌트에 해당되는지, 깊이 맵 영상 컴포넌트에 해당되는지 여부를 식별하는 임베디드 계층 영상 타입 정보, 및 상기 특정 영상 컴포넌트를 디코딩 하기 전에 디코딩되어야 하는 영상 컴포넌트의 디코딩 우선 순위를 나타내는 코딩 우선순위 정보를 더 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 1 항에 있어서,임의 어세스 포인트를 이용한 영상 컴포넌트 복원과 관련한 정보를 시그널링하는 GVR 정보 엘레먼트를 포함하는 SEI 메시지를 생성하는 단계; 및상기 GVR 정보 엘레먼트를 포함하는 SEI 메시지를 포함하는 NAL 유닛을 생성하는 단계;를 더 포함하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 6 항에 있어서, 상기 GVR 정보 엘레먼트는,특정 어세스 유닛 (access unit)이 GVR 임의 어세스 포인트인지 여부를 나타내는 GVR 식별 정보, GVR이 적용되는 영상 컴포넌트들 내에서, 현재 시점의 영상 컴포넌트 보다 시간적으로 앞서는 영상 컴포넌트가 없는 경우 디코딩이 불가능한 영상 컴포넌트의 개수를 나타내는 3D 확장 영상 개수 정보, 및 상기 디코딩이 불가능한 영상 컴포넌트를 식별하는 영상 식별 정보를 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 1 항에 있어서,상기 텍스쳐 영상 컴포넌트와 상기 깊이 맵 영상 컴포넌트 사이의 연관관계를 시그널링하는 3D 영상 의존 정보 엘레먼트를 포함하는 SEI 메시지를 생성하는 단계; 및상기 3D 영상 의존 정보 엘레먼트를 포함하는 SEI 메시지를 포함하는 NAL 유닛을 생성하는 단계;를 포함하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 8 항에 있어서, 상기 3D 영상 의존 정보 엘레먼트는,상기 3D 영상 의존 정보 엘레먼트를 포함하는 SEI 메시지가 3D 콘텐트에 포함된 모든 영상 컴포넌트들에 대한 정보를 제공하는지 여부를 나타내는 정보를 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 하나 이상의 텍스쳐 영상 컴포넌트, 상기 하나 이상의 깊이 맵 영상 컴포넌트, 및 상기 NAL 유닛을 포함하는 방송 신호를 수신하는 튜너;상기 방송 신호에서, 상기 하나 이상의 텍스쳐 영상 컴포넌트 및 하나 이상의 깊이 맵 영상 컴포넌트 사이의 코딩 관련 계층 정보를 시그널링하는 3D 계층 정보 엘레먼트를 포함하는 SEI (Supplemental Enhancement Information) 메시지를 포함하는 NAL 유닛을 파싱하는 디멀티플렉서; 및상기 SEI 메시지에 포함된 3D 계층 정보 엘레먼트를 이용하여, 상기 하나 이상의 텍스쳐 영상 컴포넌트 및 상기 하나 이상의 깊이 맵 영상 컴포넌트를 디코딩하는 영상 컴포넌트 디코더;를 포함하는 3D 영상을 위한 영상 컴포넌트 수신 장치.
- 제 10 항에 있어서, 상기 3D 계층 정보 엘레먼트는,특정 영상 컴포넌트가 텍스쳐 영상 컴포넌트에 해당되는지, 깊이 맵 영상 컴포넌트에 해당되는지 여부를 식별하는 계층 영상 타입 정보를 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 수신 장치.
- 제 11 항에 있어서, 상기 3D 계층 정보 엘레먼트는,상기 특정 영상 컴포넌트를 디코딩 하기 전에 디코딩되어야 하는 영상 컴포넌트의 개수를 나타내는 임베디드 계층 영상 개수 정보를 더 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 수신 장치.
- 제 12 항에 있어서, 상기 3D 계층 정보 엘레먼트는,상기 특정 영상 컴포넌트를 디코딩 하기 전에 디코딩되어야 하는 영상 컴포넌트가 텍스쳐 영상 컴포넌트에 해당되는지, 깊이 맵 영상 컴포넌트에 해당되는지 여부를 식별하는 임베디드 계층 영상 타입 정보, 및 상기 특정 영상 컴포넌트를 디코딩 하기 전에 디코딩되어야 하는 영상 컴포넌트의 디코딩 우선 순위를 나타내는 코딩 우선순위 정보를 더 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 수신 장치.
- 제 10 항에 있어서,상기 디멀티플렉서는,임의 어세스 포인트를 이용한 영상 컴포넌트 복원과 관련한 정보를 시그널링하는 GVR 정보 엘레먼트를 포함하는 SEI 메시지를 포함하는 NAL 유닛을 더 파싱하고,상기 영상 컴포넌트 디코더는,상기 GVR 정보 엘레먼트를 더 이용하여, 상기 하나 이상의 텍스쳐 영상 컴포넌트 및 상기 하나 이상의 깊이 맵 영상 컴포넌트를 디코딩하고,여기서, 상기 GVR 정보 엘레먼트는,특정 어세스 유닛 (access unit)이 GVR 임의 어세스 포인트인지 여부를 나타내는 GVR 식별 정보, GVR이 적용되는 영상 컴포넌트들 내에서, 현재 시점의 영상 컴포넌트 보다 시간적으로 앞서는 영상 컴포넌트가 없는 경우 디코딩이 불가능한 영상 컴포넌트의 개수를 나타내는 3D 확장 영상 개수 정보, 및 상기 디코딩이 불가능한 영상 컴포넌트를 식별하는 영상 식별 정보를 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 수신 장치.
- 제 10 항에 있어서,상기 디멀티플렉서는,상기 텍스쳐 영상 컴포넌트와 상기 깊이 맵 영상 컴포넌트 사이의 연관관계를 시그널링하는 3D 영상 의존 정보 엘레먼트를 포함하는 SEI 메시지를 포함하는 NAL 유닛을 더 파싱하고,상기 영상 컴포넌트 디코더는,상기 3D 영상 의존 정보 엘레먼트를 더 이용하여, 상기 하나 이상의 텍스쳐 영상 컴포넌트 및 상기 하나 이상의 깊이 맵 영상 컴포넌트를 디코딩하고,여기서, 상기 3D 영상 의존 정보 엘레먼트는,상기 3D 영상 의존 정보 엘레먼트를 포함하는 SEI 메시지가 3D 콘텐트에 포함된 모든 영상 컴포넌트들에 대한 정보를 제공하는지 여부를 나타내는 정보를 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 수신 장치.
- 하나 이상의 텍스쳐 영상 컴포넌트 및 하나 이상의 깊이 맵 영상 컴포넌트를 생성하는 단계;상기 하나 이상의 텍스쳐 영상 컴포넌트 및 하나 이상의 깊이 맵 영상 컴포넌트 사이의 코딩 관련 계층 정보를 시그널링하는 3D 계층 정보 디스크립터를 생성하는 단계, 여기서 상기 3D 계층 정보 디스크립터는 특정 영상 컴포넌트가 텍스쳐 영상 컴포넌트에 해당되는지, 깊이 맵 영상 컴포넌트에 해당되는지 여부를 식별하는 계층 영상 타입 정보 및 상기 특정 영상 컴포넌트를 디코딩 하기 전에 디코딩되어야 하는 영상 컴포넌트의 개수를 나타내는 임베디드 계층 영상 개수 정보를 포함하고; 및상기 하나 이상의 텍스쳐 영상 컴포넌트, 상기 하나 이상의 깊이 맵 영상 컴포넌트, 및 상기 3D 계층 정보 디스크립터를 포함하는 방송 신호를 생성하는 단계;를 포함하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 16 항에 있어서,상기 3D 계층 정보 디스크립터는, PMT, VCT, 또는 SDT에 포함되는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
- 제 16 항에 있어서, 상기 3D 계층 정보 디스크립터는,상기 특정 영상 컴포넌트를 디코딩 하기 전에 디코딩되어야 하는 영상 컴포넌트가 텍스쳐 영상 컴포넌트에 해당되는지, 깊이 맵 영상 컴포넌트에 해당되는지 여부를 식별하는 임베디드 계층 영상 타입 정보, 및 상기 특정 영상 컴포넌트를 디코딩 하기 전에 디코딩되어야 하는 영상 컴포넌트의 디코딩 우선 순위를 나타내는 코딩 우선순위 정보를 더 포함하는 것을 특징으로 하는 3D 영상을 위한 영상 컴포넌트 송신 처리 방법.
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/419,883 US9872041B2 (en) | 2012-08-10 | 2013-08-12 | Method and apparatus for transceiving image component for 3D image |
EP13827195.2A EP2884744B1 (en) | 2012-08-10 | 2013-08-12 | Method and apparatus for transceiving image component for 3d image |
CN201380042553.7A CN104541508A (zh) | 2012-08-10 | 2013-08-12 | 收发用于3d图像的图像分量的方法和装置 |
KR1020157002914A KR101653319B1 (ko) | 2012-08-10 | 2013-08-12 | 3d 영상을 위한 영상 컴포넌트 송수신 처리 방법 및 장치 |
CA2881141A CA2881141C (en) | 2012-08-10 | 2013-08-12 | Method and apparatus for transceiving image component for 3d image |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261681633P | 2012-08-10 | 2012-08-10 | |
US61/681,633 | 2012-08-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014025239A1 true WO2014025239A1 (ko) | 2014-02-13 |
Family
ID=50068393
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2013/007226 WO2014025239A1 (ko) | 2012-08-10 | 2013-08-12 | 3d 영상을 위한 영상 컴포넌트 송수신 처리 방법 및 장치 |
Country Status (6)
Country | Link |
---|---|
US (1) | US9872041B2 (ko) |
EP (1) | EP2884744B1 (ko) |
KR (1) | KR101653319B1 (ko) |
CN (1) | CN104541508A (ko) |
CA (1) | CA2881141C (ko) |
WO (1) | WO2014025239A1 (ko) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016006913A1 (ko) * | 2014-07-08 | 2016-01-14 | 엘지전자 주식회사 | 3d 방송 서비스를 위한 방송 신호 송수신 방법 및 장치 |
WO2016153273A1 (ko) * | 2015-03-23 | 2016-09-29 | 엘지전자 주식회사 | 방송 신호 송신 장치, 방송 신호 수신 장치, 방송 신호 송신 방법, 및 방송 신호 수신 방법 |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014168893A1 (en) | 2013-04-08 | 2014-10-16 | General Instrument Corporation | Signaling for addition or removal of layers in video coding |
JP2015136060A (ja) * | 2014-01-17 | 2015-07-27 | ソニー株式会社 | 通信装置、通信データ生成方法、および通信データ処理方法 |
JP6358460B2 (ja) | 2014-04-04 | 2018-07-18 | ソニー株式会社 | 受信装置、受信方法、送信装置、及び、送信方法 |
CA3060496C (en) | 2014-05-21 | 2022-02-22 | Arris Enterprises Llc | Signaling and selection for the enhancement of layers in scalable video |
MX2018013877A (es) | 2014-05-21 | 2022-06-15 | Arris Int Ip Ltd | Gestion individual de memorias intermedias en transporte de video escalable. |
TWI529662B (zh) * | 2014-12-15 | 2016-04-11 | 財團法人工業技術研究院 | 基於深度資訊的模型建構方法、圖形處理裝置與紀錄媒體 |
CN108616748A (zh) * | 2017-01-06 | 2018-10-02 | 科通环宇(北京)科技有限公司 | 一种码流及其封装方法、解码方法及装置 |
AU2020319901A1 (en) * | 2019-09-24 | 2022-04-21 | Huawei Technologies Co., Ltd. | Scalable nesting SEI messages for all layers |
WO2022144184A1 (en) * | 2020-12-28 | 2022-07-07 | Koninklijke Kpn N.V. | Prioritized decoding and output of parts of a picture in video coding |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110064146A1 (en) * | 2009-09-16 | 2011-03-17 | Qualcomm Incorporated | Media extractor tracks for file format track selection |
KR20110049137A (ko) * | 2009-11-04 | 2011-05-12 | 한국전자통신연구원 | 병렬 처리 스케일러블 비디오 부/복호화 방법 및 장치 |
KR20110111389A (ko) * | 2009-01-06 | 2011-10-11 | 엘지전자 주식회사 | 3d 비디오 신호 처리 방법 및 이와 같은 기능을 수행하는 디지털 방송 수신기 |
KR20110116239A (ko) * | 2009-02-19 | 2011-10-25 | 톰슨 라이센싱 | 3d 비디오 포맷 |
KR20120081022A (ko) * | 2009-05-01 | 2012-07-18 | 톰슨 라이센싱 | 3d 비디오 코딩 포맷 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010043773A1 (en) | 2008-10-17 | 2010-04-22 | Nokia Corporation | Sharing of motion vector in 3d video coding |
CN105120370A (zh) | 2010-12-03 | 2015-12-02 | Lg电子株式会社 | 3d广播接收器和用于接收3d广播信号的方法 |
CN102055982B (zh) | 2011-01-13 | 2012-06-27 | 浙江大学 | 三维视频编解码方法及装置 |
-
2013
- 2013-08-12 US US14/419,883 patent/US9872041B2/en active Active
- 2013-08-12 KR KR1020157002914A patent/KR101653319B1/ko active IP Right Grant
- 2013-08-12 CN CN201380042553.7A patent/CN104541508A/zh active Pending
- 2013-08-12 WO PCT/KR2013/007226 patent/WO2014025239A1/ko active Application Filing
- 2013-08-12 CA CA2881141A patent/CA2881141C/en not_active Expired - Fee Related
- 2013-08-12 EP EP13827195.2A patent/EP2884744B1/en not_active Not-in-force
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20110111389A (ko) * | 2009-01-06 | 2011-10-11 | 엘지전자 주식회사 | 3d 비디오 신호 처리 방법 및 이와 같은 기능을 수행하는 디지털 방송 수신기 |
KR20110116239A (ko) * | 2009-02-19 | 2011-10-25 | 톰슨 라이센싱 | 3d 비디오 포맷 |
KR20120081022A (ko) * | 2009-05-01 | 2012-07-18 | 톰슨 라이센싱 | 3d 비디오 코딩 포맷 |
US20110064146A1 (en) * | 2009-09-16 | 2011-03-17 | Qualcomm Incorporated | Media extractor tracks for file format track selection |
KR20110049137A (ko) * | 2009-11-04 | 2011-05-12 | 한국전자통신연구원 | 병렬 처리 스케일러블 비디오 부/복호화 방법 및 장치 |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016006913A1 (ko) * | 2014-07-08 | 2016-01-14 | 엘지전자 주식회사 | 3d 방송 서비스를 위한 방송 신호 송수신 방법 및 장치 |
KR20170010375A (ko) * | 2014-07-08 | 2017-01-31 | 엘지전자 주식회사 | 3d 방송 서비스를 위한 방송 신호 송수신 방법 및 장치 |
US9980013B2 (en) | 2014-07-08 | 2018-05-22 | Lg Electronics Inc. | Method and apparatus for transmitting and receiving broadcast signal for 3D broadcasting service |
KR101871728B1 (ko) * | 2014-07-08 | 2018-06-27 | 엘지전자 주식회사 | 3d 방송 서비스를 위한 방송 신호 송수신 방법 및 장치 |
WO2016153273A1 (ko) * | 2015-03-23 | 2016-09-29 | 엘지전자 주식회사 | 방송 신호 송신 장치, 방송 신호 수신 장치, 방송 신호 송신 방법, 및 방송 신호 수신 방법 |
US10070162B2 (en) | 2015-03-23 | 2018-09-04 | Lg Electronics Inc. | Broadcast signal transmission device, broadcast signal reception device, broadcast signal transmission method, and broadcast signal reception method |
Also Published As
Publication number | Publication date |
---|---|
US9872041B2 (en) | 2018-01-16 |
EP2884744A4 (en) | 2016-07-06 |
CA2881141C (en) | 2017-09-19 |
EP2884744A1 (en) | 2015-06-17 |
EP2884744B1 (en) | 2018-07-18 |
CN104541508A (zh) | 2015-04-22 |
KR101653319B1 (ko) | 2016-09-01 |
CA2881141A1 (en) | 2014-02-13 |
US20150229966A1 (en) | 2015-08-13 |
KR20150040902A (ko) | 2015-04-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2014025239A1 (ko) | 3d 영상을 위한 영상 컴포넌트 송수신 처리 방법 및 장치 | |
WO2010076933A1 (ko) | 이차원 영상과 3d 영상의 통합 서비스가 가능한 디지털 방송 수신방법, 및 이를 이용한 디지털 방송 수신장치 | |
WO2011013995A2 (en) | Method and apparatus for generating 3-dimensional image datastream including additional information for reproducing 3-dimensional image, and method and apparatus for receiving the 3-dimensional image datastream | |
WO2011093676A2 (en) | Method and apparatus for generating data stream for providing 3-dimensional multimedia service, and method and apparatus for receiving the data stream | |
WO2011136621A2 (en) | An apparatus of processing an image and a method of processing thereof | |
WO2010126227A2 (en) | Broadcast receiver and 3d video data processing method thereof | |
WO2011152633A2 (en) | Method and apparatus for processing and receiving digital broadcast signal for 3-dimensional subtitle | |
WO2011093677A2 (en) | Method and apparatus for transmitting digital broadcasting stream using linking information about multi-view video stream, and method and apparatus for receiving the same | |
WO2012177049A2 (en) | Method and apparatus for processing broadcast signal for 3-dimensional broadcast service | |
WO2010071283A1 (ko) | 입체영상 디스플레이가 가능한 디지털 방송 수신방법, 및 이를 이용한 디지털 방송 수신장치 | |
WO2013032221A1 (ko) | 디지털 방송 신호 처리 방법 및 장치 | |
WO2013025035A2 (ko) | 송신 장치, 수신 장치 및 그 송수신 방법 | |
WO2009145426A1 (en) | Method and apparatus for generating stereoscopic image data stream by using camera parameter, and method and apparatus for restoring stereoscopic image by using camera parameter | |
WO2010147289A1 (en) | Broadcast transmitter, broadcast receiver and 3d video processing method thereof | |
WO2012074328A2 (ko) | 다시점 3차원 방송 신호를 수신하기 위한 수신 장치 및 방법 | |
WO2011046338A2 (en) | Broadcast receiver and 3d video data processing method thereof | |
WO2010087621A2 (en) | Broadcast receiver and video data processing method thereof | |
WO2014025213A1 (ko) | 신호 송수신 장치 및 신호 송수신 방법 | |
WO2009125961A1 (en) | Method of transmitting and receiving broadcasting signal and apparatus for receiving broadcasting signal | |
WO2014204227A1 (ko) | 신호 송수신 장치 및 신호 송수신 장치의 제어 방법 | |
WO2012036532A2 (en) | Method and apparatus for processing a broadcast signal for 3d (3-dimensional) broadcast service | |
WO2012157999A2 (ko) | 비디오 스트림 전송 장치, 비디오 스트림 수신 장치, 비디오 스트림 전송 방법 및 비디오 스트림 수신 방법 | |
WO2012050405A2 (ko) | 디지털 수신기 및 디지털 수신기에서의 3d 컨텐트 처리방법 | |
WO2011028024A2 (en) | Cable broadcast receiver and 3d video data processing method thereof | |
WO2015012605A1 (ko) | 3차원 콘텐츠의 부호화 방법 및 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13827195 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20157002914 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2881141 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14419883 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2013827195 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |