WO2014073927A1 - 신호 송수신 장치 및 신호 송수신 방법 - Google Patents
신호 송수신 장치 및 신호 송수신 방법 Download PDFInfo
- Publication number
- WO2014073927A1 WO2014073927A1 PCT/KR2013/010198 KR2013010198W WO2014073927A1 WO 2014073927 A1 WO2014073927 A1 WO 2014073927A1 KR 2013010198 W KR2013010198 W KR 2013010198W WO 2014073927 A1 WO2014073927 A1 WO 2014073927A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- data
- sampling
- video data
- uhd
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 65
- 238000005070 sampling Methods 0.000 claims abstract description 189
- 230000011664 signaling Effects 0.000 claims abstract description 95
- 230000008054 signal transmission Effects 0.000 claims description 18
- 230000000750 progressive effect Effects 0.000 description 48
- 108091006146 Channels Proteins 0.000 description 25
- 238000010586 diagram Methods 0.000 description 25
- 230000001131 transforming effect Effects 0.000 description 6
- 238000006243 chemical reaction Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000012545 processing Methods 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 230000010365 information processing Effects 0.000 description 2
- PWPJGUXAGUPAHP-UHFFFAOYSA-N lufenuron Chemical compound C1=C(Cl)C(OC(F)(F)C(C(F)(F)F)F)=CC(Cl)=C1NC(=O)NC(=O)C1=C(F)C=CC=C1F PWPJGUXAGUPAHP-UHFFFAOYSA-N 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 241000023320 Luma <angiosperm> Species 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000009432 framing Methods 0.000 description 1
- OSWPMRLSEDHDFF-UHFFFAOYSA-N methyl salicylate Chemical compound COC(=O)C1=CC=CC=C1O OSWPMRLSEDHDFF-UHFFFAOYSA-N 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/266—Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
- H04N21/2662—Controlling the complexity of the video stream, e.g. by scaling the resolution or bitrate of the video stream based on the client capabilities
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
- H04N21/23439—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements for generating different versions
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/46—Colour picture communication systems
- H04N1/64—Systems for the transmission or the storage of the colour picture signal; Details therefor, e.g. coding or decoding means therefor
- H04N1/646—Transmitting or storing colour television type signals, e.g. PAL, Lab; Their conversion into additive or subtractive colour signals or vice versa therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/186—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/169—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
- H04N19/187—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scalable video layer
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
- H04N19/33—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the spatial domain
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
- H04N21/234327—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by decomposing into layers, e.g. base layer and one or more enhancement layers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/2343—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
- H04N21/234363—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by altering the spatial resolution, e.g. for clients with a lower screen resolution
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/235—Processing of additional data, e.g. scrambling of additional data or processing content descriptors
- H04N21/2353—Processing of additional data, e.g. scrambling of additional data or processing content descriptors specifically adapted to content descriptors, e.g. coding, compressing or processing of metadata
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/2362—Generation or processing of Service Information [SI]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4405—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving video stream decryption
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/63—Control signaling related to video distribution between client, server and network components; Network processes for video distribution between server and clients or between remote clients, e.g. transmitting basic layer and enhancement layers over different transmission paths, setting up a peer-to-peer communication via Internet between remote STB's; Communication protocols; Addressing
- H04N21/633—Control signals issued by server directed to the network components or client
- H04N21/6332—Control signals issued by server directed to the network components or client directed to client
- H04N21/6336—Control signals issued by server directed to the network components or client directed to client directed to decoder
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/77—Circuits for processing the brightness signal and the chrominance signal relative to each other, e.g. adjusting the phase of the brightness signal relative to the colour signal, correcting differential gain or differential phase
Definitions
- the present invention relates to a signal transmission and reception apparatus and a signal transmission and reception method.
- UHD ultra high definition
- a 4: 2: 2 or 4: 4: 4 chroma subsampling format may be used for high resolution video to provide more color.
- 4: 2: 2 format or 4: 4: 4
- the existing HDTV When using a chroma subsampling method different from the existing chroma subsampling method for high quality broadcasting, the existing HDTV also needs a method of receiving / outputting a video signal without any problem.
- the existing 4: 2: 0 chroma subsampling method and the 4: 2: 2 method are required to be compatible, and the scanning method needs to be compatible with the progressive and interlaced methods. There is a problem.
- An object of the present invention is to provide a signal transmission / reception method and a signal transmission / reception apparatus capable of providing compatibility with a subsampling format of video data.
- Another object of the present invention is to provide a signal transmitting / receiving method and a signal transmitting / receiving apparatus capable of providing a high resolution video having a different video sampling format while providing a video sampling format compatible with an existing broadcast receiver.
- Embodiments of the invention include encoding video data; Generating signaling information including video configuration metadata information capable of expressing the encoded video data in a plurality of video sampling formats; And multiplexing the encoded video data and the signaling information and transmitting the multiplexed video data and the signaling information.
- the encoded video data may include video data of a 4: 2: 2 sampling ratio of a luminance component Y and a chrominance component Cb and Cr, respectively.
- the ratio of the luminance component (Y) and the color difference components (Cb, Cr) is 4: 2: 2 sampling video data, and the ratio of the luminance component (Y) and the color difference components (Cb, Cr) is 4: 2: Video data of zero sampling, residual data of the color difference components Cb and Cr, and the video configuration metadata information may be included.
- the video configuration metadata may include the encoded video data, wherein the ratio of the luminance component (Y) to the color difference component (Cb, Cr) is 4: 2: 0 sampling of the video data, and the color difference component (Cb, Cr). It can indicate that it contains residual data.
- Another embodiment of the invention includes demultiplexing signaling information comprising video streams and video configuration metadata information indicating that the video streams can be expressed in multiple video sampling formats; Decoding the demultiplexed signaling information; And decoding video data transmitted to at least one of the video streams based on the video configuration metadata information.
- the video streams may include 4: 2: 2 sampling of video data in a ratio of luminance component Y and chrominance components Cb and Cr, respectively.
- the ratio of the luminance component (Y) and the color difference components (Cb, Cr) is 4: 2: 2 sampling video data
- the ratio of the luminance component (Y) and the color difference components (Cb, Cr) is 4: 2: Video data of zero sampling, residual data of the color difference components Cb and Cr, and the video configuration metadata information may be included.
- the video configuration metadata information may include the video streams having a ratio of 4: 2: 0 sampling of the luminance component (Y) and the chrominance component (Cb, Cr), and residuals of the chrominance component (Cb, Cr). It may indicate that it contains data.
- Another embodiment of the present invention provides an encoder for encoding video data;
- a signaling information generation unit generating signaling information including video configuration metadata information capable of expressing the encoded video data in a plurality of video sampling formats;
- a multiplexer configured to multiplex the encoded video data and the signaling information.
- Another embodiment of the present invention includes a demultiplexer for demultiplexing video streams and signaling information including video configuration metadata information indicating that the video streams can be expressed in a plurality of video sampling formats; A signaling information decoder which decodes the demultiplexed signaling information; And a decoder for decoding video data transmitted to at least one of the video streams based on the video configuration metadata information.
- video data may be transmitted and received to provide compatibility with a subsampling format of video data.
- the interlaced HD system supporting only the existing 4: 2: 0 chroma subsampling format may be provided, or high resolution video according to another chroma subsampling format may be provided.
- an HDTV receiver may receive an HD video signal through a UHDTV broadcasting channel, and a UHD receiver may also receive a signal suitable for each supported chroma subsampling format.
- FIG. 1 illustrates an embodiment of a signal transmission method according to the present invention.
- FIG. 2 is a conceptual diagram for sampling luminance (Y) and chrominance (Cb, Cr) signals in a chroma subsampling format when the video format is progressive.
- FIG. 3 is a diagram illustrating a method of encoding UHD video, which is a high resolution video
- 4 is a diagram illustrating an example of converting video data of 4: 2: 0 sampling to video data of 4: 2: 2 sampling.
- FIG. 7 is a diagram illustrating an example of a stream structure capable of providing 4: 2: 0 HD video, 4: 2: 0 UHD video, and 4: 2: 2 UHD video.
- FIG. 8 is a diagram illustrating an embodiment of a signal transmission apparatus according to an embodiment of the present invention.
- FIG. 9 is a diagram illustrating another embodiment of a signal transmission apparatus according to an embodiment of the present invention.
- FIG. 10 illustrates an example of a scanning method according to video sampling.
- FIG. 11 illustrates a first example of transforming a high resolution video into low resolution video data compatible with a scanning scheme
- FIG. 12 illustrates a second example of transforming a high resolution video into low resolution video data compatible with a scanning scheme
- FIG. 13 discloses a third example of transforming a high resolution video into low resolution video data compatible with a scanning scheme
- FIG. 14 illustrates an embodiment of a decoder of a signal receiving apparatus according to the present invention.
- FIG. 15 illustrates signaling information capable of displaying video according to an embodiment of the present invention.
- 16 is a diagram illustrating specific syntax values of signaling information according to an embodiment of the present invention.
- FIG. 17 illustrates an example of a stream level descriptor according to an embodiment of the present invention.
- 19 is a diagram illustrating information on a chroma subsampling format of original UHD video.
- 20 is a diagram illustrating resolution information on residual data of video data.
- 21 is a diagram illustrating information on a component of a signal transmitted in an enhancement layer.
- 22 is a diagram illustrating a case where illustrated descriptors are included in other signaling information.
- FIG. 23 is a diagram illustrating a case where illustrated descriptors are included in other signaling information
- 24 is a diagram illustrating a case where illustrated descriptors are included in other signaling information.
- 25 is a diagram illustrating syntax for a payload of an SEI region of video data according to embodiments.
- 26 illustrates an embodiment of a signal receiving apparatus according to the present invention.
- FIG. 27 is a view illustrating an embodiment of a signal receiving method according to the present invention
- FIG. 28 is a view illustrating an embodiment of a signal transmission apparatus according to an embodiment of the present invention.
- 29 is a view illustrating an embodiment of a signal receiving apparatus according to an embodiment of the present invention.
- FIG. 1 is a diagram illustrating an embodiment of a signal transmission method according to the present invention.
- Video data is encoded (S110).
- the video configuration metadata information UHD_composition_metadata which is encoding information of the video data, may be included in the encoded video data according to the embodiments described below.
- the encoded video data may include 4: 2: 2 sampling of video data in the ratio of the luminance component Y and the chrominance components Cb and Cr, respectively.
- the video data of the 4: 2: 2 sampling ratio of the luminance component Y and the chrominance component Cb, Cr is respectively 4:
- the ratio of the luminance component Y and the chrominance component Cb, Cr is 4:
- UHD_composition_metadata are described in detail with reference to FIG. 7 (UHD_composition_metadata) and FIG. 25.
- Video data may be scalable encoded into multiple layer data according to subsampling and resolution.
- An example of encoding video data is illustrated in FIGS. 8 and 9.
- the encoder may encode the video data so as to convert the scanning method of the video data according to the methods of FIGS. 11 to 13. Can be.
- signaling information including video configuration metadata information capable of expressing the encoded video data in a plurality of video sampling formats is generated (S120).
- the video configuration metadata information may include encoded video data in which the ratio of the luminance component (Y) to the chrominance component (Cb, Cr) is 4: 2: 0 sampling of the video data, and the residual of the chrominance component (Cb, Cr). It may indicate that it contains data. Examples of signaling information separate from the video data are illustrated in FIGS. 15 to 24.
- the encoded video data and the signaling information are multiplexed and the multiplexed video data and the signaling information are transmitted (S130). If the transmission data is not a broadcast signal, the step of generating signaling information multiplexed with the video data is omitted, and only the signaling information is encoded and transmitted together with the video data in the video data region described in S110.
- FIG. 2 illustrates a sampling conceptual diagram of luminance (Y) and color difference (Cb, Cr) signals of a chroma subsampling format for a case where a video format is progressive.
- the video format is progressive, when the luminance signal and the chrominance signal are 4: 4: 4 :, each pixel is expressed in color using both the luminance signal and the chrominance signal, and in each pixel, the ratio of the luminance signal and the chrominance signal is 1: 1: 1:
- the chrominance (Cb, Cr) signals are sampled only four times, while the luminance Y signal is sampled eight times in four pixel units of 4 x 2. The case is shown.
- the chrominance (Cb, Cr) signal is horizontally transmitted twice while the luminance Y signal is sampled eight times in four pixel units of 4 x 2. It shows the case of sampling and vertically sampling once.
- FIG. 3 is a diagram illustrating a method of encoding UHD video, which is a high resolution video.
- UHD video may be encoded using a scalable video coding method.
- a video format of 4: 2: 2 progressive format may be compatible with HD video of 4: 2: 0 interlaced format.
- the existing 4: 2: 0 interlaced HD data is encoded as base layer data.
- the base layer data may be interlaced HD data of 4: 2: 0 format.
- the Y signal may have a resolution of 1920 x 1080 and the Cb / Cr signal of 960 x 540.
- the residual luma and chroma residual signals for configuring a required 4: 2: 0 UHD signal are encoded as enhancement layer 1 data.
- the Y signal may have a resolution of 3840 x 2160 and the Cb / Cr signal of 1920 x 1080.
- Enhancement layer 2 data may be generated by encoding residual data of chroma components for constituting UHD in 4: 2: 2 format.
- the Cb / Cr signal may have a resolution of 1920 ⁇ 2160.
- the receiver merges the base layer data and the enhancement layer 1 data to display a 4: 2: 0 UHD broadcast signal, or merges the base layer data and the enhancement layer 1 and 2 data to perform a 4: 2: 2 UHD broadcast signal. Can be displayed.
- 4 illustrates an example of converting video data of 4: 2: 0 sampling into video data of 4: 2: 2 sampling.
- 4K video data of 4: 2: 0 sampling may include a luminance (Y) signal of 3840 x 2160 and a color difference (Cb, Cr) signal of 1920 x 1080, respectively.
- FIG. 5 is a diagram illustrating an upsampling process. This figure discloses an example of generating 4: 2: 2 sampling video data using enhancement layer 1 data and enhancement layer 2 data as illustrated in FIG. 3.
- the enhancement layer 2 data of the chrominance signal is residual data (r1 to r8) of the chrominance component of the high resolution video data. Therefore, enhancement layer 2 data includes pixel data having a higher resolution than enhancement layer 1 data. Therefore, 4: 2: 2 data can be generated using upscaled enhancement layer 1 color difference data and enhancement layer 2 color difference data.
- An example of this figure is a process of converting a 4: 2: 0 chroma subsampling format of 4K video to a 4: 2: 2 chroma subsampling format when the high resolution data is 4K UHD video.
- residual data of a chrominance signal (Cb / Cr) is additionally used.
- Cb / Cr residual data of a chrominance signal
- the base layer data, the enhancement layer 1 data, and the enhancement layer 2 data may be merged to finally restore the 4: 2: 2 UHD broadcast signal. .
- FIG. 6 is a diagram illustrating an upsampling process. This figure discloses an example of generating video data of 4K 4: 4: 4 chroma sampling using 4: 2: 2 chroma sample data of 4K image.
- the enhancement layer 2 data of the chrominance signal is residual data (r1 to r16) of the chrominance component of the high resolution video data. Accordingly, 4: 2: 2 sampling of high resolution video data 3840x2160 may be generated using the upscaled base layer data, enhancement layer 1 color difference data, and enhancement layer 2 color difference data.
- 0 is an example of a stream structure capable of providing HD video and 4: 2: 2 UHD video.
- a transport video stream (a) that provides 4: 2: 2 sampling of UHD video services includes streams (b), Cb residual data (e), and Cr residual data that can provide 4: 2: 0 sampling of UHD video services. (f), and metadata (g), which is signaling data.
- Stream (b) which can provide 4: 2: 0 sampling of UHD video service, provides 4: 2: 0 sampling, which can provide conventional HD video service for a receiver capable of receiving and displaying existing HD video.
- a 4: 2: 0 sampling HD video stream (c) capable of providing HD video service and residual data constituting 4: 2: 0 sampling 4K UHD video can be used to provide a 4: 2: 0 UHD video service. Can provide.
- the UHD composition metadata may include resolution information of the chrominance component and the resolution information.
- This example illustrates the case where residual data of two Chromas (Cb, Cr) is residual data of 1920 x 2160 pixel data size to provide 4K UHD video.
- the 4: 4: 4 sampling UHD video data includes: HD base layer data of 4: 2: 0 sampling, enhancement layer 1 data containing 4: 2: 0 sampling of UHD video data, and 4: 2: It may be transmitted as enhancement layer 2 data including residual data of chrominance components of 2 sampling UHD video.
- An embodiment of the apparatus for transmitting a signal may include a first encoding unit 220 for encoding base layer data, a second encoding unit 230 for encoding enhancement layer 1 data, and a third encoding unit for encoding enhancement layer 1 data. 240 may be included.
- 4K UHD original video is video (A) with 4: 4: 4 sampling. Since the original video has 4K resolution, it can be 3840 x 2160 for the Y, Cb, and Cr signals, respectively.
- the first sub sampling unit 210 may sub sample the chroma signal in the horizontal direction. Then, the UHD video B of 4: 2: 2 sampling is output. On the other hand, the original video may be 4: 2: 2 UHD video. In such a case, the first sub-sampler 210 is not included in one embodiment of the signal transmission apparatus.
- the first down sampling unit 221 of the first encoding unit 220 may down sample the video B on which the sub sampling has been performed, into a video having an HD resolution of 4: 2: 2 sampling.
- the signal converter 223 may convert the progressive video into an interlaced video. Embodiments of the conversion method will be described below with reference to the drawings.
- the second sub sampling unit 225 may sub-sample the 4: 2: 2 sampling video with respect to the chroma signal in the vertical direction.
- HD video (C) with 4: 2: 0 sampling can then be output and the resolution of this HD video can be 1920 x 1080 for luminance (Y) signals and 960 x 540 for chrominance signals (Cb, Cr).
- the first video encoder 227 may encode and output the 4: 2: 0 sampling HD video C as base layer data. Meanwhile, the positions of the signal converter 223 and the second subsampling unit 225 may be interchanged in this drawing.
- the second encoding unit 230 may include a third subsampling unit 231, an upscaling unit 233, a first calculating unit 235, and a second video encoder 237 to enhance the video data. You can encode it into data.
- the third subsampling unit 231 performs subsampling on the chroma signal in the vertical direction of the UHD video B of 4: 2: 2 sampling. Then, the UHD video D of 4: 2: 0 sampling may be output. For example, if the resolution is 4K video, the resolution of UHD video (D) with 4: 2: 0 sampling may be 3840 x 2160 for the luminance (Y) signal, or 1920 x 1080 for the chrominance signals (Cb, Cr). have.
- the upscaling unit 233 upscales and outputs the 4: 2: 0 sampling HD video C outputted by the first encoding unit 220 and outputs the UHD video size.
- the residual data of the UHD video is output by subtracting the upscaled UHD video from the: 2: 0 sampling UHD video (D).
- the second video encoder 237 may encode residual data of the UHD video and output the encoded data as enhancement layer 1 data.
- the third encoding unit 240 may include an upscaling unit 243, a second calculating unit 245, and a third video encoder 247.
- the upscaling unit 243 may upscale the chroma signal of the 4: 2: 0 sampling UHD video D.
- the second calculating unit 245 may perform the 4: 2: 2 sampling UHD video B.
- the upscaling unit 243 outputs residual video data obtained by subtracting the upscaled data of the color difference signal.
- the third video encoder 247 may encode residual video data and output the encoded video as enhancement layer 2 data.
- the third encoding unit 240 encodes video data obtained by subtracting chroma-up sampling data of 4: 2: 0 sampling UHD video D from 4: 2: 2 sampling. According to this device or its implementation, chroma subsampling can be used to transmit HD-compatible UHD video.
- down sampling of 4: 2: 2 UHD video to 4: 2: 2 HD video, conversion of progressive to interlaced, and subsampling of chroma components are performed.
- the order of the second subsampling unit 225 may be divided into three types. Three cases will be described with reference to FIGS. 10, 11, and 12.
- 4: 2: 2 UHD progressive video data is converted into 4: 2: 2 progressive HD data, and then 4: 2: 0 progressive HD data is converted to 4: 2: 0 progressive HD data.
- the following example illustrates the conversion of video to HD interlaced video.
- an example of converting to an interlaced method using chroma subsampling is disclosed.
- FIG. 12 illustrates a process of converting 4: 2: 2 UHD progressive video data into 4: 2: 0 HD interlaced video after converting 4: 2: 2 UHD progressive video data.
- the conversion is performed simultaneously with chroma subsampling and interlaced methods.
- the 4: 4: 4 sampling UHD video data includes: HD base layer data of 4: 2: 0 sampling, enhancement layer 1 data containing 4: 2: 0 sampling of UHD video data, and 4: 2: It may be transmitted as enhancement layer 2 data including residual data of chrominance components of 2 sampling UHD video.
- An embodiment of the apparatus for transmitting a signal may include a first encoder 320 for encoding base layer data, a second encoder 330 for encoding enhancement layer 1 data, and a third encoder for encoding enhancement layer 1 data. 340 may be included.
- 4K UHD original video is video (A) with 4: 4: 4 sampling. Since the original video has 4K resolution, it can be 3840 x 2160 for the Y, Cb, and Cr signals, respectively.
- the first sub sampling unit 310 may sub sample the chroma signal in the horizontal direction. Then, the UHD video B of 4: 2: 2 sampling is output. On the other hand, the original video may be 4: 2: 2 UHD video (B). In such a case, the first sub-sampler 310 is not included in the embodiment of the signal transmission apparatus.
- the second encoder 330 may include a subsampling unit 331, an upscaling unit 333, a first calculator 335, and a second video encoder 337.
- the subsampling unit 331 of the second encoder 330 performs a 4: 2: 2 sampling of UHD video B on a chroma signal in a vertical direction.
- Receive UHD video (D) of 4: 2: 0 sampling with subsampling For example, if the resolution is 4K video, the resolution of UHD video (D) with 4: 2: 0 sampling may be 3840 x 2160 for the luminance (Y) signal, or 1920 x 1080 for the chrominance signals (Cb, Cr). have.
- the signal converter 321 may convert the 4: 2: 0 sampling video from the UHD progressive method to the HD interlaced video and output the 4: 2: 0 sampling HD video C.
- the resolution of the HD video C may be 1920 x 1080 for the luminance Y signal and 960 x 540 for the chrominance signals Cb and Cr.
- the first video encoder 327 may encode and output the 4: 2: 0 sampling HD video C as base layer data.
- the second encoding unit 330 may include a subsampling unit 331, an upscaling unit 333, a first calculating unit 335, and a second video encoder 337. Enhancement can be encoded into enhancement layer 1 data.
- the subsampling unit 331 outputs the 4: 2: 0 sampling UHD video D by performing subsampling of the 4: 2: 2 sampling UHD video B on the chroma signal in the vertical direction. Can be.
- the upscaling unit 333 up-scales the 4: 2: 0 sampling HD video C outputted by the first encoding unit 320 and outputs the UHD video size, and the first calculating unit 335 is 4:
- the residual data of the UHD video is output by subtracting the upscaled UHD video from the 2: 0 sampling UHD video (D).
- the second video encoder 337 may encode residual data of the UHD video and output the encoded enhancement layer 1 data.
- the third encoder 340 may include an upscaling unit 323, a second calculator 345, and a third video encoder 347.
- the upscaling unit 343 may upscale the color difference signal of the UHD video D having 4: 2: 0 sampling.
- the second calculating unit 345 outputs residual video data obtained by subtracting upscaled data by the upscaling unit 343 from the UHD video B having 4: 2: 2 sampling.
- the third video encoder 347 may encode residual video data and output the encoded video as enhancement layer 2 data.
- the third encoding unit 340 may perform the up-sampling data of the 4: 2: 0 sampling HD video C and the residual data which is enhancement layer 1 data from the 4: 2: 2 sampling UHD video B. Encode video data subtracted from each other.
- chroma subsampling can be used to transmit HD-compatible UHD video.
- video data obtained by subsampling chroma components of a 4: 2: 2 UHD video may be converted into 4: 2: 0 interlaced video data.
- FIG. 10 is a diagram illustrating an example of a scanning method according to video sampling.
- the sample data is located in the top field of 4: 2: 2: progressive video (a) and 4: 2: 0: interlaced video (b), and in the bottom field of 4: 2: 0: interlaced video.
- Each of the videos c in which the sample data is located is illustrated.
- the chroma subsampling position may differ from field to field. Interleaving video with sampling data in the top field and video with sampling data in the bottom field is equivalent to 4: 2: 0 progressive. You can see the video in the format.
- the interlaced format compatible method can be applied to high resolution video. You can get compatible UHD video data.
- the first layer may enable 4: 2: 0 HD video
- the second layer may be 4: 2: 0 UHD video
- the third layer may be 4: 2: 2 UHD video.
- 11 discloses a first example of transforming a high resolution video into low resolution video data compatible with a scanning scheme.
- a 4: 2: 2 sampling progressive UHD video includes pixels in which both a luminance signal and a chrominance signal are located, and pixels having only a luminance signal.
- the 8 x 4 pixels included in the video will be described as an example.
- the first step converts a 4: 2: 2 sampling progressive UHD video to 4: 2: 2 sampling progressive HD video through down sampling.
- 8 x 4 pixels included in the UHD video are down sampled to 4 x 2 pixels.
- the second step is to perform down sampling on the chroma component of the 4: 2: 2 sampling progressive HD video to convert the 4: 2: 0 sampling progressive HD video.
- pixel values represented by labels indicating luminance and chrominance components of each pixel are not maintained the same at the position of the pixel during the downsampling process, but represent pixel values as a result of filtering with neighboring pixels.
- the third step converts the progressive approach to an interlaced approach. That is, 4: 2: 0 sampling progressive HD video is converted into 4: 2: 0 sampling interlaced HD video. Then only the luminance signals remain in the top field or the bottom field and the rest do not represent the pixels.
- a 4: 2: 2 progressive UHD video is converted into a 4: 2: 2 progressive HD video, which is then converted into a 4: 2: 0 progressive HD video.
- the 4: 2: 0 progressive HD video can be converted into a 4: 2: 0 interlaced HD video.
- the 4: 2: 2 sampling progressive UHD video includes pixels in which both the luminance signal and the chrominance signal are located, and pixels having only the luminance signal.
- the 8 x 4 pixels included in the video will be described as an example.
- the first step converts a 4: 2: 2 sampling progressive UHD video to 4: 2: 2 sampling progressive HD video through down sampling.
- 8 x 4 pixels included in the example UHD video are down sampled to 4 x 2 pixels.
- the second step may perform down sampling on chroma components of a 4: 2: 2 sampling progressive HD video, and simultaneously convert the progressive method into an interlaced method.
- a 4: 2: 0 interlaced HD video can be directly extracted. Then you can get 4: 2: 0 interlaced HD video.
- a 4: 2: 2 progressive UHD video is converted into a 4: 2: 2 progressive HD video.
- the 4: 2: 2 progressive HD video can be converted into a 4: 2: 0 interlaced HD video.
- FIG. 13 discloses a third example of transforming a high resolution video into low resolution video data compatible with a scanning scheme.
- the 4: 2: 2 sampling progressive UHD video includes pixels in which both the luminance signal and the chrominance signal are located, and pixels having only the luminance signal.
- the 8 x 4 pixels included in the video will be described as an example.
- the first step converts a 4: 2: 2 sampling progressive UHD video to 4: 2: 2 sampling progressive HD video through down sampling.
- 8 x 4 pixels included in the example UHD video are down sampled to 4 x 2 pixels.
- the second step downsamples the chroma components of the 4: 2: 2 sampling progressive HD video.
- the result is a 4: 2: 2 sampling interlaced HD video.
- the third step may downsample the chroma component of the 4: 2: 2 sampling interlaced HD video to obtain a 4: 2: 0 sampling interlaced HD video.
- a 4: 2: 2 progressive UHD video is converted into a 4: 2: 2 progressive HD video. Then, the 4: 2: 2 progressive HD video is converted into 4: 2: 2 interlaced HD video, which is then converted into 4: 2: 0 interlaced HD video.
- chroma component values may be assigned in consideration of coordinates when combined in a progressive manner rather than in field units.
- video data can be converted by assigning a color difference signal (eg, Cb / Cr 24) at another position using a color difference signal (Cb / Cr) with respect to a luminance signal (eg, Y16) at a specific position.
- a color difference signal eg, Cb / Cr 24
- Cb / Cr color difference signal
- a luminance signal eg, Y16
- FIG. 14 illustrates an embodiment of a decoder of a signal receiving apparatus according to the present invention as follows.
- high resolution video data is illustrated as 4K UHD video data.
- data related to video is represented by B, C and D, respectively.
- An example of a decoder that decodes video data may include at least one decoder of the base layer decoder 410, the first enhancement layer data decoder 420, and the second enhancement layer data decoder 430.
- the decoder may include a base layer decoder 410, a base layer decoder 410 and a first enhancement layer data decoder 420, or a first enhancement layer data decoder 420, And a second enhancement layer data decoder 430.
- all three functions may include a decoding unit, and the decoder of the signal receiving apparatus outputting the existing HD video may include only the base layer decoder 410.
- demultiplexing 401 may be shared by each decoding unit, and each decoding unit may include a separate demultiplexing unit 401.
- the first decoder 413 of the base layer decoder 410 may decode the base layer video data demultiplexed by the demultiplexer 401. According to the above example, the first decoder 413 may output an HD video C of 4: 2: 0 sampling. Therefore, when the transmitted video data is 4K high resolution video data, the first decoder may output video data having a resolution of 1920 ⁇ 1080 and a 960 ⁇ 540 of a chrominance signal.
- the HD receiving apparatus does not recognize Enhancement layer 1 and 2 data among the received video streams, and may recognize only HD video of the base layer and output only the HD broadcast signal even when the UHD broadcast channel is connected.
- the upscaling unit 421 of the first enhancement layer data decoder 420 may upscale the HD video decoded by the first decoder 413 into the UHD video.
- the second decoder 423 may perform scalable video decoding by using enhancement layer 1 data demultiplexed by the demultiplexer 401 and UHD video upscaled by the upscaling unit 421.
- the enhancement layer 1 data residual data of the 4: 2: 0 sampling UHD video
- the upscaling unit 421 may merge the upscaled UHD video.
- the video data decoded by the second decoder 423 may be 4: 2: 0 sampling UHD video data D.
- the second decoder 423 has a luminance signal of 3840 x 2160 and a color difference.
- the signal may output video data with a resolution of 1920 x 1080.
- a signal reception device supporting a 4: 2: 0 sampling type may output video data transmitted using base layer data and enhancement layer 1 data as UHD video.
- the upscaling unit 431 of the second enhancement layer data decoder 430 may upscale and output a chroma component of the UHD video decoded by the second decoder 423.
- the third decoder 433 may perform scalable video decoding by using enhancement layer 2 data demultiplexed by the demultiplexer 401 and UHD video upscaled by the upscaling unit 431.
- the enhancement layer 2 data residual data of the chroma component of the 4: 2: 2 sampling UHD video
- the upscaling unit 431 may merge the UHD video that upscales the chroma component.
- the video data decoded by the third decoder 433 may be 4: 2: 2 sampling UHD video data (B). Therefore, according to the example of 4K video, the third decoder 433 may output video data having a resolution of 3840 x 2160 for both the luminance signal and the color difference signal.
- the signal receiving apparatus includes at least one decoder of the above decoders, it is possible to decode and output at least one video of the HD video and the UHD video.
- Existing HD video receiving apparatus can decode only base layer data among video streams received using signaling information, and UHD video receiving apparatus outputs UHD video using base layer data and enhancement layer 1 data, or base layer data.
- the UHD video may be output using data, enhancement layer 1 data, and enhancement layer 2 data.
- signaling information of a broadcast signal capable of signaling when video is transmitted and received according to an embodiment of the present invention disclosed above is illustrated.
- FIG. 15 is a diagram illustrating signaling information capable of displaying video according to an embodiment of the present invention.
- This figure illustrates a PMT as signaling information at a system level, and may include a descriptor of a program level immediately after program_info_length of the PMT, and a descriptor of a stream level immediately after the ES_info_length field.
- descriptor_tag represents an identifier of this descriptor and descriptor_length represents a length of a descriptor.
- the UHD_program_format_type may indicate a type of service including video according to each scan type as described above.
- UHD_program_format_type is 0x05
- the service includes 4: 2: 0 sampling HD video, 4: 2: 0 sampling UHD video, and 4: 2: 2 sampling UHD video, depending on the scan type.
- a receiver capable of decoding any of these video formats may signal that it can receive and process the service.
- a UHD composition descriptor may be included in the PMT.
- This descriptor may include information about the base layer video element, information about the enhancement layer 1 video element, or information about the enhancement layer 2 video element.
- 16 is a diagram illustrating specific syntax values of signaling information according to an embodiment of the present invention.
- the illustrated field value may indicate the following information.
- the first embodiment transmits a stream for transmitting base layer data, first enhancement layer data, and second enhancement layer data, respectively, and this embodiment can signal all of these data.
- the Stream_type may be a value such as 0x02 indicating a video stream according to the MPEG-2 video codec.
- Elementary_PID represents a PID value of an elementary stream included in each program, and this example illustrates a value of 0x109A.
- the stream level descriptor may include signaling information related to MPEG-2 video.
- the PMT may include the following information.
- Stream_type may be a type value indicating a type of a stream according to the HEVC scalable layer video codec (here, illustrated as 0xA1).
- Elementary_PID represents a PID value of an elementary stream included in each program, and this example illustrates a value of 0x109B.
- UHDTV_sub_stream_descriptor which is a stream level descriptor, may include information necessary to configure 4: 2: 0 UHD video using enhancement layer 1 video data and base layer data.
- the PMT may include the following information.
- Stream_type may be a type value indicating a type of a stream according to the HEVC scalable layer video codec (here, 0xA2).
- Elementary_PID represents the PID value of the elementary stream included in each program. This example illustrates the value 0x109C.
- UHDTV_composition_descriptor which is a stream level descriptor, may include information related to enhancement layer 2 data and information required for 4: 2: 2 UHD video reconstruction.
- FIG. 17 illustrates an example of a stream level descriptor according to an embodiment of the present invention.
- the stream level descriptor may include a descriptor_tag value for identifying this descriptor, descriptor_length and UHD_composition_metadata () indicating the length of the descriptor.
- the EL2_video_codec_type field represents codec information of a video element included in the UHD service. For example, this value may have the same value as the stream_type of the PMT.
- the EL2_video_profile field may indicate profile information on the video stream, that is, information on basic specifications required for decoding the stream. Requirements information about the color depth (4: 2: 0, 4: 2: 2, etc.), bit depth (8-bit, 10-bit), coding tool, etc. of the video stream may be included. have.
- the EL2_video_level field is level information on a corresponding video stream and may include information on a technology element support range defined in a profile.
- the EL2_video_component_type field indicates what data is included when a corresponding video stream configures a UHD service. For example, does the stream transmit base layer data corresponding to 4: 2: 0 sampling HD video, or enhancement layer 1 data, which is residual data included in 4: 2: 0 sampling UHD video, or 4: Indicates whether enhancement layer 2 data included in a 2: 2 sampling UHD video is transmitted.
- the original_UHD_video_type field signals information about the UHD video format and may represent basic information such as a video resolution and a frame rate. This will be described in detail later.
- original_UHD_video_chroma_subsampling_type Indicates information about the chroma subsampling format of the original UHD video. This field will be described later.
- the EL2_video_resolution field indicates resolution information about residual data to be added, which will be described later.
- the EL2_video_type field represents information about a component of a signal transmitted by the enhancement layer, and this information will be described later.
- the chroma_upscale_filter field represents information related to a filter when upscaling a chroma component from a first scan type to a second scan type. For example, it provides filter related information specifying chroma upscaling method from 4: 2: 0 scan type to 4: 2: 2 scan type.
- the filter related information may include filter tap information or filter coefficient information.
- the original_UHD_video_type field of the signaling information may indicate the resolution and frame rate of the video. This figure illustrates that various resolutions and frame rates may be provided according to the value. For example, if the original_UHD_video_type field is 0101, the original video may have 60 frames per second and a 4096x2160 resolution.
- the original_UHD_video_chroma_subsampling_type field of the described signaling information indicates information about a chroma subsampling format of the original UHD video. This figure illustrates, for example, that when this value is 10, the subsampling format of the luminance signal and the color difference signal is 4: 4: 4 (Y: Cb: Cr).
- the EL2_video_resolution field of the described signaling information indicates resolution information about residual data to be added. For example, when the EL2_video_resolution field is 010, this indicates that the resolution of the residual data of the high resolution chrominance component, which is enhancement layer 2 data constituting the video, is 1920 x 2160.
- the EL2_video_type field of the described signaling information indicates information about a component of a signal transmitted by the enhancement layer. For example, when the EL2_video_type field is 01, this illustrates that a signal component of enhancement layer 2 data constituting video is a Cb signal.
- FIG. 22 illustrates a case where the above-described descriptors are included in other signaling information. This figure illustrates the case where the illustrated descriptors are included in the SDT.
- the table_id field represents an identifier of a table.
- section_syntax_indicator field is a 1-bit field set to 1 for an SDT table section (section_syntax_indicator: The section_syntax_indicator is a 1-bit field which shall be set to "1")
- section_length This is a 12-bit field, the first two bits of which shall be "00" .It specifies the number of bytes of the section, starting immediately following the section_length field and including the CRC. The section_length shall not exceed 1 021 so that the entire section has a maximum length of 1 024 bytes.
- transport_stream_id This is a 16-bit field which serves as a label for identification of the TS, about which the SDT informs, from any other multiplex within the delivery system.
- the version_number field indicates the version number of this subtable.
- version_number This 5-bit field is the version number of the sub_table.
- the version_number shall be incremented by 1 when a change in the information carried within the sub_table occurs.When it reaches value "31", it wraps around to "0" .
- the version_number shall be that of the currently applicable sub_table.
- the version_number shall be that of the next applicable sub_table.
- section_number This 8-bit field gives the number of the section.
- the section_number of the first section in the sub_table shall be "0x00" .
- the section_number shall be incremented by 1 with each additional section with the same table_id, transport_stream_id, and original_network_id.
- last_section_number This 8-bit field specifies the number of the last section (that is, the section with the highest section_number) of the sub_table of which this section is part.
- service_id This is a 16-bit field which serves as a label to identify this service from any other service within the TS.
- the service_id is the same as the program_number in the corresponding program_map_section.
- EIT_schedule_flag This is a 1-bit field which when set to "1" indicates that EIT schedule information for the service is present in the current TS , see TR 101 211 [i.2] for information on maximum time interval between occurrences of an EIT schedule sub_table). If the flag is set to 0 then the EIT schedule information for the service should not be present in the TS.
- the EIT_present_following_flag field may indicate whether EIT_present_following_flag: This is a 1-bit field which when set to "1" indicates that EIT_present_following information for the service is present in the current TS, see TR 101 211 [i.2] for information on maximum time interval between occurrences of an EIT present / following sub_table.If the flag is set to 0 then the EIT present / following information for the service should not be present in the TS.)
- the running_status field may refer to the state of the service defined in Table 6 of the DVB-SI document.
- running_status This is a 3-bit field indicating the status of the service as defined in table 6.For an NVOD reference service the value of the running_status shall be set to "0".
- free_CA_mode This 1-bit field, when set to “0" indicates that all the component streams of the service are not scrambled.When set to "1" it indicates that access to one or more streams may be controlled by a CA system.
- descriptors_loop_length field indicates the length of the following descriptor (descriptors_loop_length: This 12-bit field gives the total length in bytes of the following descriptors).
- CRC_32 This is a 32-bit field that contains the CRC value that gives a zero output of the registers in the decoder
- the descriptors_loop_length field may include the UHD_program_type_descriptor illustrated in FIG. 15 and the UHD_composition_descriptor illustrated in FIG. 17 in the following descriptor position.
- the UHD_component_descriptor may further include a component_tag field.
- the component_tag field may indicate a PID value for a corresponding stream signaled by PMT, which is a PSI level.
- the receiver may find the PID value of the corresponding stream together with the PMT using the component_tag field.
- FIG. 23 illustrates a case where the illustrated descriptors are included in other signaling information. This figure illustrates the case where the illustrated descriptors are included in the EIT.
- EIT may be in accordance with ETSI EN 300 468. Using this to describe each field is as follows.
- table_id Represents a table identifier.
- section_syntax_indicator field is a 1-bit field set to 1 for an EIT table section (section_syntax_indicator: The section_syntax_indicator is a 1-bit field which shall be set to "1").
- section_length This is a 12-bit field.It specifies the number of bytes of the section, starting immediately following the section_length field and including the CRC.The section_length shall not exceed 4 093 so that the entire section has a maximum length of 4 096 bytes.
- service_id This is a 16-bit field which serves as a label to identify this service from any other service within a TS.
- the service_id is the same as the program_number in the corresponding program_map_section.
- the version_number field indicates the version number of this subtable.
- version_number This 5-bit field is the version number of the sub_table.
- the version_number shall be incremented by 1 when a change in the information carried within the sub_table occurs.When it reaches value 31, it wraps around to 0.When the current_next_indicator is set to "1”, then the version_number shall be that of the currently applicable sub_table.When the current_next_indicator is set to "0”, then the version_number shall be that of the next applicable sub_table.
- section_number This 8-bit field gives the number of the section.
- the section_number of the first section in the sub_table shall be "0x00" .
- the section_number shall be incremented by 1 with each additional section with the same table_id, service_id, transport_stream_id, and original_network_id.
- the sub_table may be structured as a number of segments.With each segment the section_number shall increment by 1 with each additional section, but a gap in numbering is permitted between the last section of a segment and the first section of the adjacent segment.
- last_section_number This 8-bit field specifies the number of the last section (that is, the section with the highest section_number) of the sub_table of which this section is part.
- transport_stream_id This is a 16-bit field which serves as a label for identification of the TS, about which the EIT informs, from any other multiplex within the delivery system.
- segment_last_section_number This 8-bit field specifies the number of the last section of this segment of the sub_table.For sub_tables which are not segmented, this field shall be set to the same value as the last_section_number field.
- the last_table_id field is (last_table_id: This 8-bit field identifies the last table_id used (see table 2).)
- event_id This 16-bit field contains the identification number of the described event (uniquely allocated within a service definition).
- the start_time field contains the start time of the event (start_time: This 40-bit field contains the start time of the event in Universal Time, Co-ordinated (UTC) and Modified Julian Date (MJD) (see annex C). is coded as 16 bits giving the 16 LSBs of MJD followed by 24 bits coded as 6 digits in 4-bit Binary Coded Decimal (BCD) . If the start time is undefined (eg for an event in a NVOD reference service) all bits of the field are set to "1".)
- running_status This is a 3-bit field indicating the status of the event as defined in table 6. For an NVOD reference event the value of the running_status shall be set to "0".
- free_CA_mode This 1-bit field, when set to “0" indicates that all the component streams of the event are not scrambled.When set to “1” it indicates that access to one or more streams is controlled by a CA system.
- descriptors_loop_length field indicates the length of the following descriptor. (descriptors_loop_length: This 12-bit field gives the total length in bytes of the following descriptors.)
- CRC_32 This is a 32-bit field that contains the CRC value that gives a zero output of the registers in the decoder
- the descriptors_loop_length field may include the UHD_program_type_descriptor illustrated in FIG. 15 and the UHD_composition_descriptor illustrated in FIG. 17 in the following descriptor position.
- the UHD_component_descriptor may further include a component_tag field.
- the component_tag field may indicate a PID value for a corresponding stream signaled by PMT, which is a PSI level.
- the receiver may find the PID value of the corresponding stream together with the PMT using the component_tag field.
- FIG. 24 illustrates a case where the above-described descriptors are included in other signaling information. This figure illustrates the case where the illustrated descriptors are included in the VCT.
- the VCT may comply with the ATSC PSIP specification. According to ATSC PSIP, each field is described as follows. Each bit description starts as follows.
- the section_syntax_indicator field is a 1-bit field set to 1 for a VCT table section (section_syntax_indicator-The section_syntax_indicator is a one-bit field which shall be set to '1' for the terrestrial_virtual_channel_table_section ()).
- section_length field represents the length of the section in bytes. (section_length-This is a twelve bit field, the first two bits of which shall be ‘00’. It specifies the number of bytes of the section, starting immediately following the section_length field, and including the CRC.)
- the transport_stream_id field indicates an MPEG-TS ID as in a PAT that can identify TVCT (transport_stream_id-The 16-bit MPEG-2 Transport Stream ID, as it appears in the Program Association Table (PAT) identified by a PID value of zero for this multiplex.
- the transport_stream_id distinguishes this Terrestrial Virtual Channel Table from others that may be broadcast in different PTCs.
- the version_number field indicates the version number of the VCT (version_number-This 5 bit field is the version number of the Virtual Channel Table.
- version number shall be incremented by 1 whenever the definition of the current VCT changes. Upon reaching the value 31, it wraps around to 0.
- version number shall be one unit more than that of the current VCT (also in modulo 32 arithmetic) In any case, the value of the version_number shall be identical to that of the corresponding entries in the MGT)
- current_next_indicator-A one-bit indicator, which when set to '1' indicates that the Virtual Channel Table sent is currently applicable.When the bit is set to ' 0 ', it indicates that the table sent is not yet applicable and shall be the next table to become valid. This standard imposes no requirement that “next” tables (those with current_next_indicator set to' 0 ') must be sent.An update to the currently applicable table shall be signaled by incrementing the version_number field)
- section_number-This 8 bit field gives the number of this section.
- the section_number of the first section in the Terrestrial Virtual Channel Table shall be 0x00. It shall be incremented by one with each additional section in the Terrestrial Virtual Channel Table)
- last_section_number-This 8 bit field specifies the number of the last section (that is, the section with the highest section_number) of the complete Terrestrial Virtual Channel Table.)
- protocol_version field indicates the protocol version for a parameter to be defined differently from the current protocol (protocol_version-An 8-bit unsigned integer field whose function is to allow, in the future, this table type to carry parameters that may be structured differently than those defined in the current protocol.At present, the only valid value for protocol_version is zero.Non-zero values of protocol_version may be used by a future version of this standard to indicate structurally different tables)
- the num_channels_in_section-This 8 bit field specifies the number of virtual channels in this VCT section. The number is limited by the section length)
- the major_channel_number field indicates the number of major channels associated with a virtual channel (major_channel_number-A 10-bit number that represents the “major” channel number associated with the virtual channel being defined in this iteration of the “for” loop.Each virtual channel shall be associated with a major and a minor channel number. The major channel number, along with the minor channel number, act as the user's reference number for the virtual channel. The major_channel_number shall be between 1 and 99. The value of major_channel_number shall be set such that in no case is a major_channel_number / minor_channel_number pair duplicated within the TVCT.For major_channel_number assignments in the US, refer to Annex B.)
- the minor_channel_number field indicates the number of minor channels associated with the virtual channel (minor_channel_number-A 10-bit number in the range 0 to 999 that represents the "minor" or "sub"-channel number.This field, together with major_channel_number, performs as a two-part channel number, where minor_channel_number represents the second or right-hand part of the number.When the service_type is analog television, minor_channel_number shall be set to 0.
- Minor_channel_number shall be set such that in no case is a major_channel_number / minor_channel_number pair duplicated within the TVCT.For other types of services, such as data broadcasting, valid minor virtual channel numbers are between 1 and 999.
- modulation_mode mode indicates the modulation mode of the carrier associated with the virtual channel (modulation_mode-An 8-bit unsigned integer number that indicates the modulation mode for the transmitted carrier associated with this virtual channel.Values of modulation_mode shall be as defined in Table 6.5. digital signals, the standard values for modulation mode (values below 0x80) indicate transport framing structure, channel coding, interleaving, channel modulation, forward error correction, symbol rate, and other transmission-related parameters, by means of a reference to an appropriate standard The modulation_mode field shall be disregarded for inactive channels)
- carrier_frequency-The recommended value for these 32 bits is zero.Use of this field to identify carrier frequency is allowed, but is deprecated.
- channel_TSID field indicates the MPEG-2 TS ID associated with the TS carrying the MPEG-2 program referenced by this virtual channel (channel_TSID-A 16-bit unsigned integer field in the range 0x0000 to 0xFFFF that represents the MPEG-2 Transport Stream ID associated with the Transport Stream carrying the MPEG-2 program referenced by this virtual channel8.
- channel_TSID shall represent the ID of the Transport Stream that will carry the service when it becomes active. The receiver is expected to use the channel_TSID to verify that any received Transport Stream is actually the desired multiplex.
- channel_TSID shall indicate the value of the analog TSID included in the VBI of the NTSC signal.Refer to Annex D Section 9 for a discussion on use of the analog TSID)
- the program_number field indicates an integer value defined in association with this virtual channel and PMT (program_number-A 16-bit unsigned integer number that associates the virtual channel being defined here with the MPEG-2 PROGRAM ASSOCIATION and TS PROGRAM MAP tables.
- PMT program_number-A 16-bit unsigned integer number that associates the virtual channel being defined here with the MPEG-2 PROGRAM ASSOCIATION and TS PROGRAM MAP tables.
- channels representing analog services a value of 0xFFFF shall be specified for program_number.
- program_number shall be set to zero.This number shall not be interpreted as pointing to a Program Map Table entry .
- the access_controlled field may refer to an event associated with an access controlled virtual channel (access_controlled-A 1-bit Boolean flag that indicates, when set, that the events associated with this virtual channel may be access controlled.When the flag is set to '0', event access is not restricted)
- the hidden field may indicate that the virtual channel is not accessed by the user's direct channel input (hidden-A 1-bit Boolean flag that indicates, when set, that the virtual channel is not accessed by the user by direct entry of the virtual channel number.Hidden virtual channels are skipped when the user is channel surfing, and appear as if undefined, if accessed by direct channel entry.Typical applications for hidden channels are test signals and NVOD services. Whether a hidden channel and its events may appear in EPG displays depends on the state of the hide_guide bit.)
- the hide_guide field may indicate whether a virtual channel and its events may be displayed in the EPG (hide_guide-A Boolean flag that indicates, when set to '0' for a hidden channel, that the virtual channel and its events may appear in EPG displays.This bit shall be ignored for channels which do not have the hidden bit set, so that non-hidden channels and their events may always be included in EPG displays regardless of the state of the hide_guide bit.
- Typical applications for hidden channels with the hide_guide bit set to '1' are test signals and services accessible through application-level pointers.
- service_type-This 6-bit field shall carry the Service Type identifier.Service Type and the associated service_type field are defined in A / 53 Part 1 [1] to identify the type of service carried in this virtual channel.Value 0x00 shall be reserved.Value 0x01 shall represent analog television programming.Other values are defined in A / 53 Part 3 [3], and other ATSC Standards may define other Service Types9)
- the source_id field is an identification number identifying a program source associated with a virtual channel (source_id-A 16-bit unsigned integer number that identifies the programming source associated with the virtual channel.
- a source is one specific source of video, text Source ID values zero is reserved.Source ID values in the range 0x0001 to 0x0FFF shall be unique within the Transport Stream that carries the VCT, while values 0x1000 to 0xFFFF shall be unique at the regional level.Values for source_ids 0x1000 and above shall be issued and administered by a Registration Authority designated by the ATSC.)
- the descriptors_length field indicates the length of the descriptor following (descriptors_length-Total length (in bytes) of the descriptors for this virtual channel that follows)
- descriptor ()-Zero or more descriptors, as appropriate, may be included.
- the service_type field may have a field value indicating parameterized service (0x07) or extended parameterized service (0x09) or scalable UHDTV service.
- the UHD_program_type_descriptor illustrated in FIG. 15 and the UHD_composition_descriptor illustrated in FIG. 17 may be located at the descriptor position.
- 25 illustrates syntax for a payload of an SEI region of video data according to embodiments of the present invention.
- payloadType When payloadType is set to a specific value (52 in this example) in the SEI payload, it may include information (UHD_scalable_chroma__serivce_info (payloadSize)) signaling the format of the video data as illustrated.
- this signaling information represents signaling information related to the scalable coded chroma component.
- An embodiment of parsing video data according to syntax illustrated by a decoder of a receiver is as follows.
- the decoder When the decoder decodes the video data, it parses the AVC or HEVC NAL unit from the video element stream. If the nal_unit_type value corresponds to SEI data and the payloadType is 52 in the SEI data, information according to the syntax of the UHDTV_scalable_chroma_service_info SEI message can be obtained.
- the syntax of the illustrated SEI payload may indicate program format type information (UHD_program_format_type) indicating that video data is transmitted to a plurality of layers according to sampling information, as illustrated in FIG. 14.
- UHD_program_format_type program format type information
- the video data may include video configuration metadata information UHD_composition_metadata including sampling information of the video data illustrated in FIG. 17.
- the receiver can determine whether the decoded video data is 4: 2: 0 HD video data, 4: 2: 0 UHD video data, or 4: 2: 2 UHD video data. Based on this information, the receiver can output UHD or HD video that matches the receiver's performance.
- 26 is a diagram illustrating an embodiment of a signal receiving apparatus according to the present invention.
- the signal receiving apparatus may include a demultiplexer 500, a signaling information decoder 600, and a video decoder 700.
- the demultiplexer 500 may demultiplex video streams and signaling information according to an embodiment of the present invention.
- the video streams may include the streams illustrated in FIG. 7.
- the signaling information processing unit 600 may decode a part of the signaling information processing unit 600 according to the signaling information illustrated in FIGS. 15 to 24 or the performance of the receiver.
- the video decoder 700 may recognize that video data is coded into a plurality of layer data according to sampling information based on program format type information (UHD_program_format_type) in the signaling information.
- the video decoder 700 may include at least one video decoder of the first decoder 710, the second decoder 720, and the third decoder 730 according to the performance of the receiver.
- the video decoder 700 includes a first decoder 710, includes a first decoder 710 and a second decoder 720, or a first decoder 710, a second decoder 720. And a third decoder 730.
- the first decoder 710 receives a stream for transmitting base layer data based on the video configuration metadata information (UHD_composition_metadata) of the signaling information, decodes the base layer data, and performs 4: 2: 0 sampling of video data (C). ) Can be decoded and output.
- the base layer data may include video configuration metadata information (UHD_composition_metadata) of the video data in the video data region illustrated in FIG. 25, and the first decoder 710 decodes the base layer data accordingly to 4: 2: 0. Video data C of sampling can be output.
- the first upscaler 715 may upscale and output HD video data of 4: 2: 0 sampling, which is base layer data, to a high resolution (UHD) video with respect to a luminance (Y) component and a color difference (Cb, Cr) component. have.
- the second decoder 720 includes residual data of 4: 2: 0 UHD video, which is enhancement layer 1, based on configuration metadata information (UHD_composition_metadata) of the signaling information, and HD video upscaled from the first upscaler 715.
- 4: 2: 0 UHD video data can be output.
- Residual data of 4: 2: 0 UHD video which is an enhancement layer 1, may include configuration metadata information (UHD_composition_metadata) of video data in the video data region illustrated in FIG. 25, and the second decoder 720 may thus set 4. : 2: 0 UHD video data D can be decoded.
- configuration metadata information UHD_composition_metadata
- the second upscaler 725 may upscale and output a color difference (Cb, Cr) component of 4: 2: 0 sampling UHD video data, which is enhancement layer 1 data.
- the third decoder 730 may include residual data of the color difference (Cb, Cr) component of the 4: 2: 2 UHD video, which is an enhancement layer 2, based on the configuration metadata information (UHD_composition_metadata) of the signaling information, and the second upscaler ( From 725, the 4: 2: 2 UHD video data B may be output by decoding the upscaled UHD video data having the color difference Cb and Cr components.
- Cb, Cr color difference
- UHD_composition_metadata configuration metadata information
- FIG. 27 is a view illustrating an embodiment of a signal receiving method according to the present invention.
- An embodiment of a signal receiving method demultiplexes signaling information including video streams and video configuration metadata information indicating that the video streams can be expressed in multiple video sampling formats (S210). ).
- the signaling information may be demultiplexed only when the received signal is a broadcast signal.
- the received video streams may include 4: 2: 2 sampling of video data in a ratio of the luminance component Y and the chrominance components Cb and Cr, respectively.
- the video data of 4: 2: 2 sampling ratio of the luminance component Y and the said chrominance component Cb, Cr is respectively 4: 2:
- the ratio of the luminance component Y and the said chrominance component Cb, Cr is 4: 2:
- Video data of zero sampling, residual data of the color difference components Cb and Cr, and video configuration metadata information may be included.
- the structure of the video transmitted in the received video stream may be in accordance with the structure illustrated in FIG. 7. That is, the transport video stream (a) that provides 4: 2: 2 sampling of UHD video service includes the stream (b), Cb residual data (e), and Cr that can provide 4: 2: 0 sampling of UHD video service. residual data f, and metadata g, which is signaling data.
- Stream (b), which can provide 4: 2: 0 sampling of UHD video service provides 4: 2: 0 sampling, which can provide conventional HD video service for a receiver capable of receiving and displaying existing HD video.
- HD video stream (c) and enhancement layer data (d) for 4: 2: 0 sampling 4K UHD video data.
- the signaling information illustrated in FIGS. 15 to 24 may be demultiplexed separately from the video data.
- the demultiplexed signaling information may be decoded (S220). If the received signal is not a broadcast signal, step S220 may be omitted and information decoded signaling information in the video data may be used in the following video data decoding step.
- the video configuration metadata information includes a video stream having a ratio of 4: 2: 0 sampling of the luminance component (Y) to the chrominance component (Cb, Cr) and residual data of the chrominance component (Cb, Cr). May be included.
- FIGS. 15 to 24 are illustrated.
- the video data transmitted to at least one of the video streams is decoded based on the video configuration metadata information (S230).
- the video data decoding process according to the performance of the receiver and the video configuration metadata information decoded as illustrated in FIG. 26 results in 4: 2: 0 HD video data, 4: 2: 0 UHD video data, Alternatively, 4: 2: 2 UHD video data can be decoded.
- the process of decoding the video data has been described with reference to FIG. 14 or 26.
- FIG. 28 is a view illustrating an embodiment of a signal transmission apparatus according to an embodiment of the present invention.
- An embodiment of the apparatus for transmitting a signal may include an encoder 810, a signaling information generator 820, and a multiplexer 830.
- Encoder 810 encodes the video data.
- the video configuration metadata information UHD_composition_metadata which is encoding information of the video data, may be included in the encoded video data according to the disclosed embodiment.
- the encoded video data may include video data of a 4: 2: 2 sampling ratio of the luminance component Y and the color difference components Cb and Cr, respectively, and the luminance component Y and the color difference component Cb,
- the video data of 4: 2: 2 sampling of the ratio of Cr) is the video data of the 4: 2: 0 sampling of the ratio of the luminance component (Y) and the color difference component (Cb, Cr), and the color difference component (Cb, Cr). Residual data, and the video configuration metadata information.
- Video configuration metadata information (UHD_composition_metadata) of encoded video data is described in detail with reference to FIG. 7 (UHD_composition_metadata) and FIG. 25.
- the encoder 810 may be scalable encoded with multiple layer data according to subsampling and resolution. An example of encoding video data is illustrated in FIGS. 8 and 9.
- the scanning method of the video data may be converted according to the methods of FIGS. 11 to 13.
- an embodiment of the apparatus for transmitting a signal includes a signaling information generator 820 separately from the encoder 810.
- the signaling information generation unit 820 generates signaling information including video configuration metadata information that enables the encoder 810 to express encoded video data in a plurality of video sampling formats. Examples of signaling information separate from the video data are illustrated in FIGS. 15 to 24.
- the video configuration metadata information includes the encoded video data having a ratio of 4: 2: 0 sampling of the luminance component (Y) to the chrominance component (Cb, Cr) and the chrominance component (Cb, Cr). It can indicate that it contains residual data.
- the multiplexer 830 may multiplex encoded video data and signaling information and output multiplexed video data and signaling information. If the transmission data is not a broadcast signal, the signaling information generator 820 for generating signaling information multiplexed with the video data is omitted, and the multiplexer 830 stores the video component meta in the video data region encoded by the encoder 810. Video data including data information (UHD_composition_metadata) and other data (for example, audio data) are multiplexed and output.
- UHD_composition_metadata data information
- other data for example, audio data
- an example of the signal transmission apparatus may include 4: 2: 0 sampling HD video, 4: 2 to display video data according to a receiver performance and a subsampling type of video data.
- Signaling information for selectively configuring: 0 UHD video or 4: 2: 2 UHD video and video data according to the structure may be transmitted.
- 29 is a view illustrating an embodiment of a signal receiving apparatus according to an embodiment of the present invention.
- One embodiment of the apparatus for receiving a signal may include a demultiplexer 910, a signaling information decoder 920, and a video decoder 930.
- the demultiplexer 910 demultiplexes the video stream and the signaling information.
- the signaling information is demultiplexed only when the received signal is a broadcast signal.
- the structure of the demultiplexed video streams is illustrated in FIG. 7. As illustrated in FIG. 7, video configuration metadata information of video data may be included in the demultiplexed video stream, and the syntax thereof is described in detail with reference to FIG. 25.
- the video streams may comprise 4: 2: 2 sampling of video data in a ratio of the luminance component Y and the chrominance components Cb and Cr, respectively.
- the video data of the 4: 2: 2 sampling ratio of the luminance component (Y) and the chrominance component (Cb, Cr), respectively, has a 4: 2: 0 sampling ratio of the luminance component (Y) and the chrominance component (Cb, Cr).
- Video data, residual data of the color difference components Cb and Cr, and the video configuration metadata information may be included.
- the demultiplexer 910 demultiplexes the signaling information of FIGS. 15 to 24 transmitted separately from the video data, and when the received video stream is not a broadcast signal, the signaling information is illustrated in FIG. 7. It may be included in the video configuration metadata information.
- the signaling information decoder 920 decodes the demultiplexed signaling information.
- the demultiplexed signaling information may include the information illustrated in FIGS. 15 to 24, and the information illustrated in the above drawings may be decoded according to an embodiment. If the received signal is not a broadcast signal, the signaling information decoder 920 may be omitted and may use information obtained by decoding signaling information in the video data in the following video data decoding step.
- the video composition metadata information includes video data in which the ratios of the luminance component (Y) and the chrominance component (Cb, Cr) are 4: 2: 0 sampling, and the residual data of the chrominance component (Cb, Cr). May indicate that it is.
- the video decoder 930 decodes the video data in the video data area or according to the video configuration metadata information that is the signaling information decoded by the signaling information decoding unit 920.
- a detailed example of the video decoder 930 has been described above with reference to FIG. 26.
- 4: 2: 0 sampling HD video, 4: 2: 0 UHD video, or 4: 2: 2 UHD video may be selected according to the performance of the receiver and the subsampling type of video data. Can output selectively.
- the transmitting apparatus transmits the structure of video data as shown in FIG. 7
- information about the structure may be obtained from the signaling information and the video data may be decoded and output according to the performance of the receiver.
- the present invention has industrial applicability that is usable and repeatable in the fields of broadcast and video signal processing.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- Library & Information Science (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Color Television Systems (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
Claims (15)
- 비디오 데이터를 인코딩하는 단계;상기 인코딩된 비디오 데이터를, 다수의 비디오 샘플링 포맷으로 표출할 수 있는 비디오 구성 메타데이터 정보를 포함하는 시그널링 정보를 생성하는 단계; 및상기 인코딩된 비디오 데이터와 상기 시그널링 정보를 다중화하고 상기 다중화된 비디오 데이터와 시그널링 정보를 전송하는 단계를 포함하는, 신호 송신 방법.
- 제 1항에 있어서,상기 인코딩된 비디오 데이터는, 휘도 성분(Y)과 색차 성분(Cb, Cr)의 비가 각각 4:2:2 샘플링의 비디오 데이터를 포함하는 신호 송신 방법.
- 제 2항에 있어서,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 각각 4:2:2 샘플링의 비디오 데이터는,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 4:2:0 샘플링의 비디오 데이터, 상기 색차 성분(Cb, Cr)의 residual 데이터, 및 상기 비디오 구성 메타데이터 정보를 포함하는, 신호 송신 방법.
- 제 1항에 있어서,상기 비디오 구성 메타데이터 정보는,상기 인코딩된 비디오 데이터가,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 4:2:0 샘플링의 비디오 데이터, 상기 색차 성분(Cb, Cr)의 residual 데이터를 포함하고 있음을 나타내는, 신호 송신 방법.
- 비디오 스트림들과, 상기 비디오 스트림들이 다수의 비디오 샘플링 포맷으로 표출될 수 있음을 나타내는 비디오 구성 메타데이터 정보를 포함하는 시그널링 정보를 역다중화하는 단계;상기 역다중화한 시그널링 정보를 디코딩하는 단계; 및상기 비디오 구성 메타데이터 정보에 기초하여 상기 비디오 스트림들 중 적어도 하나에 전송되는 비디오데이터를 디코딩하는 단계를 포함하는, 신호 수신 방법.
- 제 5항에 있어서,상기 비디오 스트림들은,휘도 성분(Y)과 색차 성분(Cb, Cr)의 비가 각각 4:2:2 샘플링의 비디오 데이터를 포함하는, 신호 수신 방법.
- 제 6항에 있어서,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 각각 4:2:2 샘플링의 비디오 데이터는,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 4:2:0 샘플링의 비디오 데이터, 상기 색차 성분(Cb, Cr)의 residual 데이터, 및 상기 비디오 구성 메타데이터 정보를 포함하는, 신호 수신 방법.
- 제 5항에 있어서,상기 비디오 구성 메타데이터 정보는,상기 비디오 스트림들이,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 4:2:0 샘플링의 비디오 데이터, 상기 색차 성분(Cb, Cr)의 residual 데이터를 포함하고 있음을 나타내는, 신호 수신 방법.
- 비디오 데이터를 인코딩하는 인코더;상기 인코딩된 비디오 데이터를, 다수의 비디오 샘플링 포맷으로 표출할 수 있는 비디오 구성 메타데이터 정보를 포함하는 시그널링 정보를 생성하는 시그널링정보생성부; 및상기 인코딩된 비디오 데이터와 상기 시그널링 정보를 다중화하는 다중화부;를 포함하는, 신호 송신 장치.
- 제 9항에 있어서,상기 인코딩된 비디오 데이터는, 휘도 성분(Y)과 색차 성분(Cb, Cr)의 비가 각각 4:2:2 샘플링의 비디오 데이터를 포함하는 신호 송신 장치.
- 제 10항에 있어서,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 각각 4:2:2 샘플링의 비디오 데이터는,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 4:2:0 샘플링의 비디오 데이터, 상기 색차 성분(Cb, Cr)의 residual 데이터, 및 상기 비디오 구성 메타데이터 정보를 포함하는, 신호 송신 장치.
- 제 9항에 있어서,상기 비디오 구성 메타데이터 정보는,상기 인코딩된 비디오 데이터가,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 4:2:0 샘플링의 비디오 데이터, 상기 색차 성분(Cb, Cr)의 residual 데이터를 포함하고 있음을 나타내는, 신호 송신 장치.
- 비디오 스트림들과, 상기 비디오 스트림들이 다수의 비디오 샘플링 포맷으로 표출될 수 있음을 나타내는 비디오 구성 메타데이터 정보를 포함하는 시그널링 정보를 역다중화하는 역다중화부;상기 역다중화한 시그널링 정보를 디코당하는 시그널링정보복호부; 및상기 비디오 구성 메타데이터 정보에 기초하여 상기 비디오 스트림들 중 적어도 하나에 전송되는 비디오데이터를 디코딩하는 디코더;를 포함하는, 신호 수신 장치.
- 제 13항에 있어서,상기 비디오 스트림들은,휘도 성분(Y)과 색차 성분(Cb, Cr)의 비가 각각 4:2:2 샘플링의 비디오 데이터를 포함하는, 신호 수신 장치.
- 제 13항에 있어서,상기 비디오 구성 메타데이터 정보는,상기 비디오 스트림들이,상기 휘도 성분(Y)과 상기 색차 성분(Cb, Cr)의 비가 4:2:0 샘플링의 비디오 데이터, 상기 색차 성분(Cb, Cr)의 residual 데이터를 포함하고 있음을 나타내는 신호 수신 장치.
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/441,443 US9554162B2 (en) | 2012-11-12 | 2013-11-11 | Apparatus for transreceiving signals and method for transreceiving signals |
CA2890508A CA2890508C (en) | 2012-11-12 | 2013-11-11 | Apparatus for transreceiving signals and method for transreceiving signals |
CN201380058983.8A CN104782135B (zh) | 2012-11-12 | 2013-11-11 | 用于收发信号的设备和用于收发信号的方法 |
JP2015541693A JP6126234B2 (ja) | 2012-11-12 | 2013-11-11 | 信号送受信装置及び信号送受信方法 |
EP13853021.7A EP2919471A4 (en) | 2012-11-12 | 2013-11-11 | SIGNAL TRANSMITTING APPARATUS AND SIGNAL TRANSMITTING-RECEIVING METHOD |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261725472P | 2012-11-12 | 2012-11-12 | |
US61/725,472 | 2012-11-12 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014073927A1 true WO2014073927A1 (ko) | 2014-05-15 |
Family
ID=50684940
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/KR2013/010198 WO2014073927A1 (ko) | 2012-11-12 | 2013-11-11 | 신호 송수신 장치 및 신호 송수신 방법 |
Country Status (7)
Country | Link |
---|---|
US (1) | US9554162B2 (ko) |
EP (1) | EP2919471A4 (ko) |
JP (1) | JP6126234B2 (ko) |
KR (1) | KR20150087209A (ko) |
CN (1) | CN104782135B (ko) |
CA (1) | CA2890508C (ko) |
WO (1) | WO2014073927A1 (ko) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016208832A1 (ko) * | 2015-06-22 | 2016-12-29 | 주식회사 디엠티 | 에이치디 셋톱박스용 유에이치디 어댑터 및 유에이치디 티브이, 유에이치 티브이에 연결하기 위한 에이치디 셋톱박스 |
Families Citing this family (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
MX352400B (es) * | 2013-07-12 | 2017-11-23 | Sony Corp | Dispositivo de reproducción, método de reproducción y medio de grabación. |
WO2015013645A1 (en) | 2013-07-25 | 2015-01-29 | Convida Wireless, Llc | End-to-end m2m service layer sessions |
US20160241906A1 (en) * | 2013-09-16 | 2016-08-18 | Samsung Electronics Co., Ltd. | Broadcasting reception apparatus and control method |
JP2017507547A (ja) * | 2014-01-03 | 2017-03-16 | トムソン ライセンシングThomson Licensing | ビデオコンテンツをレンダリングするときの超高精細解像度へのアップスケーリングを最適化するための方法、装置、およびコンピュータプログラム製品 |
US9998765B2 (en) | 2014-07-16 | 2018-06-12 | Qualcomm Incorporated | Transport stream for carriage of video coding extensions |
KR101699337B1 (ko) * | 2014-08-20 | 2017-01-24 | 전자부품연구원 | Shvc 기반의 uhd 영상데이터 송수신 시스템 |
CN115209186A (zh) | 2014-10-10 | 2022-10-18 | 索尼公司 | 再现装置和再现方法 |
US10462502B2 (en) * | 2014-11-21 | 2019-10-29 | Sony Corporation | Transmitting apparatus, transmitting method, receiving apparatus, and receiving method |
WO2017180439A1 (en) * | 2016-04-15 | 2017-10-19 | Vid Scale, Inc. | System and method for fast stream switching with crop and upscale in client player |
US10956766B2 (en) | 2016-05-13 | 2021-03-23 | Vid Scale, Inc. | Bit depth remapping based on viewing parameters |
WO2017211306A1 (zh) * | 2016-06-08 | 2017-12-14 | 同济大学 | 视频数据压缩码流的解码、视频数据的编码方法及装置 |
US10771795B2 (en) * | 2016-06-13 | 2020-09-08 | Intel Corporation | Switchable chroma sampling for wireless display |
EP3482566B1 (en) | 2016-07-08 | 2024-02-28 | InterDigital Madison Patent Holdings, SAS | Systems and methods for region-of-interest tone remapping |
CN109891772B (zh) | 2016-11-03 | 2022-10-04 | 康维达无线有限责任公司 | Nr中的帧结构 |
WO2018152437A1 (en) | 2017-02-17 | 2018-08-23 | Vid Scale, Inc. | Systems and methods for selective object-of-interest zooming in streaming video |
US11178204B1 (en) * | 2017-02-23 | 2021-11-16 | Cox Communications, Inc. | Video processor to enhance color space and/or bit-depth |
KR102628139B1 (ko) | 2017-03-07 | 2024-01-22 | 인터디지털 매디슨 페턴트 홀딩스 에스에이에스 | 멀티-디바이스 프리젠테이션을 위한 맞춤형 비디오 스트리밍 |
WO2020068251A1 (en) | 2018-09-27 | 2020-04-02 | Convida Wireless, Llc | Sub-band operations in unlicensed spectrums of new radio |
CN117956169A (zh) * | 2022-10-31 | 2024-04-30 | 华为技术有限公司 | 图像编解码方法和装置 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR19990081409A (ko) * | 1998-04-29 | 1999-11-15 | 구자홍 | 디지탈 티브이 수신 디코더 장치 |
KR20000038809A (ko) * | 1998-12-09 | 2000-07-05 | 윤종용 | 디스플레이 포맷 및/또는 프레임 레이트가 서로 다른 신호들을동시에 표시하는 수신기와 그 표시방법 |
US20020030620A1 (en) * | 2000-09-05 | 2002-03-14 | Cairns Graham Andrew | Multi-format active matrix displays |
KR20040048702A (ko) * | 2002-12-04 | 2004-06-10 | 엘지전자 주식회사 | 디지털 비디오 포맷을 지원하는 엠펙 비디오 디코딩 시스템 |
KR20120019751A (ko) * | 2010-08-26 | 2012-03-07 | 에이스텔 주식회사 | 초고화질 영상을 위한 실시간 3d 포맷팅 모듈 및 이를 이용한 실시간 3d 포맷팅 시스템 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050129130A1 (en) * | 2003-12-10 | 2005-06-16 | Microsoft Corporation | Color space coding framework |
EP1737240A3 (en) * | 2005-06-21 | 2007-03-14 | Thomson Licensing | Method for scalable image coding or decoding |
KR20080114388A (ko) * | 2007-06-27 | 2008-12-31 | 삼성전자주식회사 | 스케일러블 영상 부호화장치 및 방법과 그 영상 복호화장치및 방법 |
JP5242111B2 (ja) * | 2007-10-02 | 2013-07-24 | 株式会社ソニー・コンピュータエンタテインメント | 送信装置、画像データ送信方法、受信装置および受信装置における画像表示方法 |
KR101213704B1 (ko) * | 2007-12-05 | 2012-12-18 | 삼성전자주식회사 | 가변 컬러 포맷 기반 동영상 부호화 방법 및 장치, 그복호화 방법 및 장치 |
KR20090097015A (ko) * | 2008-03-10 | 2009-09-15 | 삼성전자주식회사 | 스케일러블 영상 부호화장치 및 스케일러블 영상복호화장치 |
JP2010268092A (ja) * | 2009-05-13 | 2010-11-25 | Sony Corp | 送信装置および送信方法、受信装置および受信方法、並びにプログラム |
CA2765570C (en) * | 2009-06-16 | 2016-05-03 | Lg Electronics Inc. | Broadcast transmitter, broadcast receiver and 3d video processing method thereof |
-
2013
- 2013-11-11 WO PCT/KR2013/010198 patent/WO2014073927A1/ko active Application Filing
- 2013-11-11 CA CA2890508A patent/CA2890508C/en not_active Expired - Fee Related
- 2013-11-11 EP EP13853021.7A patent/EP2919471A4/en not_active Ceased
- 2013-11-11 KR KR1020157011734A patent/KR20150087209A/ko not_active Application Discontinuation
- 2013-11-11 JP JP2015541693A patent/JP6126234B2/ja active Active
- 2013-11-11 US US14/441,443 patent/US9554162B2/en not_active Expired - Fee Related
- 2013-11-11 CN CN201380058983.8A patent/CN104782135B/zh not_active Expired - Fee Related
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR19990081409A (ko) * | 1998-04-29 | 1999-11-15 | 구자홍 | 디지탈 티브이 수신 디코더 장치 |
KR20000038809A (ko) * | 1998-12-09 | 2000-07-05 | 윤종용 | 디스플레이 포맷 및/또는 프레임 레이트가 서로 다른 신호들을동시에 표시하는 수신기와 그 표시방법 |
US20020030620A1 (en) * | 2000-09-05 | 2002-03-14 | Cairns Graham Andrew | Multi-format active matrix displays |
KR20040048702A (ko) * | 2002-12-04 | 2004-06-10 | 엘지전자 주식회사 | 디지털 비디오 포맷을 지원하는 엠펙 비디오 디코딩 시스템 |
KR20120019751A (ko) * | 2010-08-26 | 2012-03-07 | 에이스텔 주식회사 | 초고화질 영상을 위한 실시간 3d 포맷팅 모듈 및 이를 이용한 실시간 3d 포맷팅 시스템 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2919471A4 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2016208832A1 (ko) * | 2015-06-22 | 2016-12-29 | 주식회사 디엠티 | 에이치디 셋톱박스용 유에이치디 어댑터 및 유에이치디 티브이, 유에이치 티브이에 연결하기 위한 에이치디 셋톱박스 |
KR101728076B1 (ko) * | 2015-06-22 | 2017-04-19 | 주식회사 디엠티 | 에이치디 셋톱박스용 유에이치디 어댑터 및 유에이치디 티브이, 유에이치 티브이에 연결하기 위한 에이치디 셋톱박스 |
Also Published As
Publication number | Publication date |
---|---|
US20150304693A1 (en) | 2015-10-22 |
CA2890508C (en) | 2017-08-15 |
CA2890508A1 (en) | 2014-05-15 |
JP2016501458A (ja) | 2016-01-18 |
EP2919471A1 (en) | 2015-09-16 |
KR20150087209A (ko) | 2015-07-29 |
CN104782135A (zh) | 2015-07-15 |
JP6126234B2 (ja) | 2017-05-10 |
US9554162B2 (en) | 2017-01-24 |
EP2919471A4 (en) | 2016-07-13 |
CN104782135B (zh) | 2018-02-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2014073927A1 (ko) | 신호 송수신 장치 및 신호 송수신 방법 | |
WO2014084564A1 (ko) | 신호 송수신 장치 및 신호 송수신 방법 | |
WO2014025213A1 (ko) | 신호 송수신 장치 및 신호 송수신 방법 | |
WO2015076616A1 (ko) | 신호 송수신 장치 및 신호 송수신 방법 | |
WO2015152635A1 (ko) | 신호 송수신 장치 및 신호 송수신 방법 | |
WO2015126144A1 (ko) | 파노라마 서비스를 위한 방송 신호 송수신 방법 및 장치 | |
WO2014073853A1 (ko) | 신호 송수신 장치 및 신호 송수신 방법 | |
WO2015072754A1 (ko) | Hdr 방송 서비스 제공을 위한 방송 신호 송수신 방법 및 장치 | |
WO2015102449A1 (ko) | 컬러 개멋 리샘플링을 기반으로 하는 방송 신호 송수신 방법 및 장치 | |
WO2014077651A1 (ko) | 신호 송수신 장치 및 신호 송수신 방법 | |
WO2014003379A1 (ko) | 영상 디코딩 방법 및 이를 이용하는 장치 | |
WO2010071283A1 (ko) | 입체영상 디스플레이가 가능한 디지털 방송 수신방법, 및 이를 이용한 디지털 방송 수신장치 | |
WO2009125961A1 (en) | Method of transmitting and receiving broadcasting signal and apparatus for receiving broadcasting signal | |
WO2011093677A2 (en) | Method and apparatus for transmitting digital broadcasting stream using linking information about multi-view video stream, and method and apparatus for receiving the same | |
WO2012023789A2 (ko) | 디지털 방송 신호 수신 장치 및 방법 | |
WO2015065037A1 (ko) | Hevc 기반의 ip 방송 서비스 제공을 위한 방송 신호 송수신 방법 및 장치 | |
WO2014104725A1 (ko) | 영상의 부호화/복호화 방법 및 이를 이용하는 장치 | |
WO2014038906A1 (ko) | 영상 복호화 방법 및 이를 이용하는 장치 | |
WO2015080414A1 (ko) | 트릭 플레이 서비스 제공을 위한 방송 신호 송수신 방법 및 장치 | |
WO2015034306A1 (ko) | 디지털 방송 시스템에서 고화질 uhd 방송 컨텐츠 송수신 방법 및 장치 | |
WO2012050405A2 (ko) | 디지털 수신기 및 디지털 수신기에서의 3d 컨텐트 처리방법 | |
WO2014107069A1 (ko) | 영상 부호화/복호화 방법 및 장치 | |
WO2015093811A1 (ko) | 트릭 플레이 서비스 제공을 위한 신호 송수신 장치 및 신호 송수신 방법 | |
WO2015115869A1 (ko) | 트릭 플레이 서비스 제공을 위한 신호 송수신 장치 및 신호 송수신 방법 | |
WO2013162256A1 (ko) | 3d 서비스를 위한 신호 처리 장치 및 방법 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13853021 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20157011734 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2890508 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2013853021 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14441443 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2015541693 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |