US20110012992A1 - Simulcast of stereoviews for 3d tv - Google Patents

Simulcast of stereoviews for 3d tv Download PDF

Info

Publication number
US20110012992A1
US20110012992A1 US12/835,103 US83510310A US2011012992A1 US 20110012992 A1 US20110012992 A1 US 20110012992A1 US 83510310 A US83510310 A US 83510310A US 2011012992 A1 US2011012992 A1 US 2011012992A1
Authority
US
United States
Prior art keywords
video
transport stream
independently compressed
descriptor
base
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US12/835,103
Other versions
US9036700B2 (en
Inventor
Ajay K. Luthra
Mandayam A. Narasimhan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Google Technology Holdings LLC
Original Assignee
General Instrument Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to US12/835,103 priority Critical patent/US9036700B2/en
Application filed by General Instrument Corp filed Critical General Instrument Corp
Priority to EP10800518.2A priority patent/EP2454886B1/en
Priority to PCT/US2010/042072 priority patent/WO2011008917A1/en
Priority to CN201080031847.6A priority patent/CN102474635B/en
Priority to KR1020127003916A priority patent/KR101342294B1/en
Assigned to GENERAL INSTRUMENT CORPORATION reassignment GENERAL INSTRUMENT CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LUTHRA, AJAY K., NARASIMHAN, MANDAYAM A.
Publication of US20110012992A1 publication Critical patent/US20110012992A1/en
Assigned to GENERAL INSTRUMENT HOLDINGS, INC. reassignment GENERAL INSTRUMENT HOLDINGS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GENERAL INSTRUMENT CORPORATION
Assigned to MOTOROLA MOBILITY LLC reassignment MOTOROLA MOBILITY LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GENERAL INSTRUMENT HOLDINGS, INC.
Assigned to Google Technology Holdings LLC reassignment Google Technology Holdings LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MOTOROLA MOBILITY LLC
Application granted granted Critical
Publication of US9036700B2 publication Critical patent/US9036700B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/161Encoding, multiplexing or demultiplexing different image signal components
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

Definitions

  • Depth perception for a three dimensional (3D) TV is provided by capturing two views, one for the left eye and other for the right eye. These two views are compressed and sent over various networks or stored on storage media.
  • a decoder decodes the two views and sends the decoded video to the 3D TV for display.
  • the two views are known to be either merged into a single video frame or kept separate. When kept separate, the two views are known to be compressed and distributed using various methods. Two of these methods are multi-view coding (MVC) method and simulcast method.
  • MVC multi-view coding
  • simulcast method the two views are compressed such that a correlation between the two views is exploited to increase coding efficiency.
  • the two views are known to be compressed and distributed as two independent video streams.
  • MPEG standards such as advanced video coding (AVC) or H.264 MPEG-4 and MPEG-2 systems
  • AVC advanced video coding
  • MPEG-4 and MPEG-2 systems provide methods and syntax to represent and compress more than one view as well as to transport this compressed data in various ways.
  • the current versions of the MPEG standards do not specify how to simulcast two views so that the two views can decoded and combined appropriately to provide a 3D display. Consequently, simulcasting may not be currently done using MPEG standards for video systems and other types of systems.
  • broadcast centric standards i.e. Society of Cable Telecommunications Engineers (SCTE) and Advanced Televisions Systems Committee (ATSC)
  • SCTE Society of Cable Telecommunications Engineers
  • ATSC Advanced Televisions Systems Committee
  • This constraint precludes simulcasting of two video streams corresponding to the two eye views using a same compression standard in a same program.
  • the constraint precludes simulcasting two independent MPEG-2 or MPEG-4 streams (i.e. one for the left eye and one for the right eye for 3D TV) in the same program.
  • Sending the two views in separate programs may be possible but adds complexity of requiring an additional tuner and time base synchronization between the two programs.
  • both eye views are compressed using the same encoding standard. Because there are two dominant encoding standards, MPEG-2 and MPEG-4, this creates a backward compatibility problem for simulcast methods in digital TV systems that have mixture of MPEG-2 and MPEG-4 decoders. For instance, the MPEG-2 decoders will not be able to decode either view if the two views are compressed using the MPEG-4 standard.
  • a three-dimensional (3D) video encoder includes a first encoder, a second encoder, and a multiplexer.
  • the first encoder is configured to encode a base view to form a first independently compressed video stream associated with video content.
  • the second encoder is configured to encode a nonbase view to form a second independently compressed video stream associated with the video content.
  • the multiplexer is configured to multiplex the first and second independently compressed video streams associated with the video content to form a transport stream.
  • the transport stream is operable to be processed to render a 3D video using the base and nonbase views.
  • a 3D video decoder includes an input module, a decoding module and a 3D combining module.
  • the input module is configured to receive a transport stream.
  • the transport stream includes first and second independently compressed video streams associated with the video content, included in a same broadcast program in the transport stream.
  • the decoding module is configured to decode the transport stream to form base and nonbase views that are respectively from the first and second independently compressed video streams associated with the video content.
  • the 3D combining module is configured to render a 3D video using the base and nonbase views.
  • a method of encoding a transport stream includes receiving video content.
  • a base view is thereafter encoded to form a first independently compressed video stream associated with the video content.
  • a nonbase view is encoded to form a second independently compressed video stream associated with the video content.
  • the first and second independently compressed video streams associated with the video content are multiplexed to form a transport stream.
  • the transport stream is operable to be processed to render a 3D video using the base and nonbase views.
  • a method of rendering a transport stream into 3D video includes receiving the transport stream, which includes first and second independently compressed video streams associated with the video content, included in a same broadcast program in the transport stream.
  • the transport stream is then decoded to form base and nonbase views that are respectively formed from the first and second independently compressed video streams associated with the video content.
  • a 3D video is thereafter rendered using the base and nonbase views.
  • a computer readable storage medium on which is embedded one or more computer programs implements the above-disclosed method of rendering a transport stream into 3D video.
  • Embodiments of the present invention include simulcasting independently compressed video streams in a same program for 3D TV distribution systems.
  • the compressed video streams may have different standards, such as MPEG 2 and MPEG 4 to allow for backward compatibility.
  • the compressed video streams may have the same standards, such as both MPEG 4, to provide the highest resolution and quality 3D image in systems that have compatible decoders.
  • the independently compressed video streams are simulcasted in the same program by sending the streams as different types. By sending the streams in the same program rather than in two different programs, synchronization between the streams is much easier for generating the 3D video.
  • FIG. 1 illustrates a network architecture, according to an embodiment of the invention
  • FIG. 2 illustrates a functional block diagram of multiplexing first and second independently compressed video streams associated with the video content to form a transport stream, according to an embodiment of the invention
  • FIG. 3 illustrates a simplified block diagram of a 3D video decoder, according to an embodiment of the invention
  • FIG. 4 illustrates a flow diagram of a method of encoding a transport stream, according to an embodiment of the invention
  • FIG. 5 illustrates a flow diagram of a method of rendering a transport stream into 3D video, according to an embodiment of the invention, according to an embodiment of the invention
  • FIG. 6 illustrates a flow diagram of a method of rendering a transport stream into 3D video, according to an embodiment of the invention.
  • FIG. 7 shows a block diagram of a computer system that may be used in the 3D video decoder, according to an embodiment of the invention.
  • FIG. 1 illustrates a network architecture 100 of a system in which a 3D video decoder 130 may be used, according an embodiment.
  • the network architecture 100 is illustrated as a cable television (CATV) network architecture, including a cable head-end unit 110 and a cable network 111 .
  • a number of data sources 101 , 102 , 103 may be communicatively coupled to the cable head-end unit 110 including, but in no way limited to, a plurality of servers 101 , the Internet 102 , radio signals, or television signals received via a content provider 103 .
  • the cable head-end 110 is also communicatively coupled to one or more subscribers 150 a - n through the cable network 111 .
  • the network architecture 100 depicted in FIG. 1 may include additional components and that some of the components described herein may be removed and/or modified without departing from a scope of the network architecture 100 .
  • the cable head-end 110 is configured to output a transport stream 131 .
  • the transport stream 131 may be output to the set top boxes (STB) 120 a - n using a satellite transmission, the Internet or other network using, for instance, fixed optical fibers or coaxial cables.
  • the transport stream 131 is configured to provide cross standard compatibility for a mixture of 3D video decoders and 2D decoders 121 that may be used by the STBs 120 a - n .
  • Each of the subscribers 150 a - n may receive the transport stream 131 at STBs 120 a - n .
  • the STBs 120 a - n are devices that enable TVs to receive digital television (DTV) broadcasts.
  • the STBs 120 a - n includes any device, not just conventional STBs that can perform the functions of the STBs 120 a - n described herein.
  • the STBs 120 a - n may be equipped with decoders with either 2D capability, as described with respect to the 2D decoder 121 or 3D capability.
  • the STBs 120 a - n may be connected to either a 2D TV 123 or a 3D-ready TV 122 .
  • the STB with 3D capability may be connected to a 3D-ready TV 122 .
  • the 3D video decoder 130 and an STB 120 c in which it is located may be used in other networks and that the cable network architecture 100 is shown for simplicity and illustrative purposes.
  • the 3D video decoder 130 may be used in other architectures.
  • FIG. 2 illustrates a functional block diagram of a 3D video encoder 200 .
  • the 3D video encoder 200 includes a first encoder 202 , a second encoder 204 , a descriptor processor 208 and a multiplexer 206 .
  • the 3D video encoder 200 depicted in FIG. 2 may include additional components and that some of the components described herein may be removed and/or modified without departing from a scope of the 3D video encoder 200 .
  • the first encoder 202 is configured to compress a base view to form a first independently compressed video stream 132 a associated with video content.
  • a base view is an eye view for the 3D video that may be used as a full view in a two dimensional (2D) system.
  • the base view may comprise a left eye view in a 3D video system and the full view in a 2D video system.
  • the video content may be received from, for instance, broadcast programs, Internet Protocol TV (IPTV), switched video (SDV), video on demand (VOD) or other video sources.
  • IPTV Internet Protocol TV
  • SDV switched video
  • VOD video on demand
  • the second encoder is configured to compress a nonbase view to form a second independently compressed video stream 132 b associated with the video content.
  • a nonbase view is used along with the base view to render the 3D video, as described hereinbelow, and may be dropped by 2D systems.
  • the first and second independently compressed video streams 132 a - b associated with the video content are not compressed with reference to other components, e.g. in contrast to MVC coded video components in which the components are coded with respect to each other.
  • Example of video compression may include conventional MPEG standards.
  • MPEG refers to a group of standards for encoding and decoding transport streams as defined by the Moving Picture Experts Group. MPEG includes several formats for transport streams such as MPEG-2 and MPEG-4 part 10/H.264.
  • the descriptor processor 208 is configured to create and add an associated descriptor 209 to the first and second independently compressed video streams 132 a - b .
  • the descriptor processor 208 may receive information describing the first and second independently compressed video streams 132 a - b from the first encoder 202 and the second encoder 204 respectively, and use this information in generating the descriptor 209 .
  • the descriptor 209 is configured to allow a decoder that receives the transport stream 131 to decode (including any adjustment of video resolutions) and render a 3D video from the decoded base and nonbase views and is included in both the first and second independently compressed video streams 132 a - b.
  • the multiplexer 206 is configured to receive the first and second independently compressed video streams associated with the video content 132 a - b from the first encoder 202 and the second encoder 204 . The multiplexer 206 thereafter multiplexes the first and second independently compressed video streams 132 a - b along with the associated descriptor 209 to form a transport stream 131 .
  • the transport stream 131 is a video stream that includes the first and second independently compressed video streams 132 a - b .
  • the independently compressed video streams 132 a - b may be included in a same program in the transport stream 131 . A same program may correspond, for instance, to a single channel in broadcasting.
  • the transport stream 131 may comprise Internet Protocol (IP) packets or any packetization scheme configured to transmit the transport stream 131 .
  • IP Internet Protocol
  • the transport stream 131 may comprise an MPEG stream, IP packets or any suitable medium for transporting the 3D video from a source to an end user.
  • the base view in the first compression format is a full resolution 2D compatible stream that may be received and decoded by a 2D decoder, for instance the 2D decoder 121 shown in FIG. 1 that receives the same transport stream 131 .
  • the nonbase view may be discarded by the 2D decoder 121 , for instance a conventional 2D decoder.
  • the transport stream 131 may be used to provide backwards compatibility in a network architecture having a mixture of STBs with conventional 2D decoders and STBs with 3D decoders the network architecture 100 with the conventional decoders while allowing 3D capability with the STB 120 c.
  • the first independently compressed video stream 132 a is configured in a first compression format, for instance MPEG-2.
  • the second independently compressed video stream 132 b is configured in a second compression format, which differs from the first compression format, for instance MPEG-4.
  • the broadcast-centric constraints allow two video components with different stream_type values within same program (i.e., the existing stream_type values for MPEG-2 video (0x02) and MPEG-4 (0x1B) can be used).
  • the first and second independently compressed video streams 132 a - b are configured in a common compression format, for instance both being MPEG-2 or both being MPEG-4.
  • the second independently compressed video stream 132 b may be at a lower resolution than the first independently compressed video stream 132 a .
  • the second independently compressed video stream 132 b is assigned a new stream type that is different than the first independently compressed video stream 132 a in the program map table for the transport stream 131 .
  • a program map table is used to describe each single program including elementary streams associated with the program.
  • the stream type may be specified in the program map table.
  • the new stream type for the nonbase view may be allocated by a transport system, for instance MPEG-2.
  • the new stream type allocated by the MPEG-2 systems may be equal to 0x21 for MPEG-4 and equal to 0x22 for MPEG-2. Alternately, the new stream type may be determined using a next available stream type, for instance a next available ‘reserved’ stream type values in the MPEG-2 systems (International Organization for Standardization (ISO) 13818-1) stream type table.
  • ISO International Organization for Standardization
  • the two views compressed independently may be sent (using the same compression standard such as MPEG-4) as part of the same broadcast program in the transport stream 131 if sent as two different stream types.
  • the transport stream 131 also includes a descriptor 209 that provides information to assist in combining and rendering the two separately decoded images onto an appropriate display.
  • the descriptor 209 provides information to assist in combining and rendering two separately decoded images from the base view and the nonbase view into a 3D image to be displayed at the 3D-ready TV 122 .
  • the descriptor 209 may specify that the independently compressed video stream is for a particular view. Thus, even though the steam type for the video stream may be reserved or some other stream type, the decoder, for instance the 3D decoder 130 in FIG.
  • the descriptor 209 can determine from the descriptor 209 that the stream is to be combined with another stream for the other view.
  • the descriptor 209 may also include identification of view (e.g. left or right), upsampling rules in instances where view resolutions differ, information for combining the views (e.g. field interleaving or frame interleaving) and additional display related elements.
  • the descriptor 209 may be configured to provide information on use of single stream, combining and rendering two fields or two frames in a same video sequence and additional display related elements in instances in which the transport stream 131 is an MPEG-4 conformant stream.
  • the single stream is used in contrast to two streams to carry the base and nonbase views.
  • the information provided in the descriptor 209 at a system level includes a 3D rendering information descriptor.
  • the 3D rendering information descriptor includes a descriptor tag, a descriptor length and an instruction to render each picture in the 3D video.
  • An example of the 3D rendering information descriptor in tabular form is the following:
  • the 3D rendering information descriptor provides information such as field_parity, top_field_first_in_composition, frame_order and display_specific_info including left_or_right_view, up_sample_horizontal, up_sample_vertical and field_or_frame_interleaving. For instance, by determining the descriptor constant k, the decoder determines whether the 3D video is to be rendered top field first in composition.
  • the display specific information may include resolution of the 3D video 222 .
  • the 3D rendering information descriptor may be used by a receiver, for instance the STB 120 c , to render the received the transport stream 131 to form 3D video 222 .
  • FIG. 3 illustrates a simplified block diagram of a system 300 configured to receive and display 3D content, according to an embodiment.
  • the system 300 comprises the STB 120 c , including the 3D decoder 130 , and the 3D-ready TV 122 from FIG. 1 .
  • the STB 120 c includes an input/output module 210 , a demultiplexing module 212 , and the 3D video encoder 130 , which includes a 3D decoding module 214 , and a 3D combining module 220 .
  • the system 300 depicted in FIG. 3 may include additional components and that some of the components described herein may be removed and/or modified without departing from a scope of the system 300 .
  • the input/output module 210 is configured to receive the transport stream 131 .
  • the input/output module 210 may comprise a Universal Serial Bus (USB), an Ethernet interface, or another type of interface.
  • USB Universal Serial Bus
  • the transport stream 131 received at the input/output module 210 in this instance may be a QAM modulated stream.
  • the demultiplexing module 212 may thereafter select a frequency and demodulate the frequency to obtain a multi program transport stream (MPTS).
  • MPTS multi program transport stream
  • the demultiplexing module 212 is configured to demultiplex the MPTS to extract single program transport streams (SPTSs) corresponding to each of a plurality of programs which a subscriber may select. For instance, the subscriber 150 n may use the STB 120 to select a program having 3D content.
  • the demultiplexing module 212 then demultiplexes the MPTS to form the first and second independently compressed video streams associated with the video content 132 a - b . Each of the first and second independently compressed video streams associated with the video content 132 a - b has a different codec.
  • the 3D decoding module 214 is configured to decode the first and second independently compressed video streams 132 a - b to form a base view (2D view) corresponding to a first eye view provided and a nonbase view corresponding to a second eye view.
  • the base view and the nonbase view may be used by the STB 120 c to render a 3D video 222 and output for display at the 3D-ready TV 122 .
  • the 3D combining module 220 may be configured to access the 3D rendering information descriptor in the transport stream 131 and render the 3D video 222 using the instruction therein. After decoding of the first and second independently compressed video streams associated with the video content 132 a - b by the decoding module 214 , the base and nonbase views may be combined by the 3D combining module 220 to form a 3D video 222 that may be displayed on a 3D-ready TV 122 capable of displaying 3D (stereoscopic or multiple) views.
  • the 3D combining module 220 may use a descriptor 209 included in the transport stream 131 in combining the base and nonbase views to render the 3D video 222 .
  • the descriptor 209 may include the supplemental enhancement information (SEI) to form a 3D video 222 .
  • SEI Supplemental Enhancement Information
  • the input/output module 210 is configured to thereafter output the 3D video 222 , for instance to a 3D-ready TV.
  • the 3D combining module 220 may thereafter render pictures in the base view and the nonbase view according to information provided in the 3D rendering information descriptor.
  • the 3D rendering information descriptor specifies operations to be performed to render the 3D video 222 from the transport stream 131 . For instance, if two views are received in the same stream, the 3D combining module 220 may determine whether field pictures are received. Similarly, if field pictures are received in the base view, the 3D combining module 220 renders the 3D video 222 from the transport stream 131 according to the syntax in Table 1, hereinabove.
  • some of the rendering and combining information from the descriptor 209 may be included in the compressed video stream (such as picture user-data). This provides information on use of single stream, combining and rendering two fields or two frames in the same video sequence and additional display related elements.
  • the MPEG-4 video based rendering and combining information from the descriptor 209 includes an SEI message providing such information at a picture level or slice level in which values represent picture size, base view picture size, Sample Aspect Ratio (SAR), number of legs in the re-sampling poly-phase filter in the horizontal direction (Lh), number of filter coefficients (N) in each of those legs, number of legs in the re-sampling poly-phase filter in the vertical direction (Lv) and number of filter coefficients (M) in each of those legs.
  • SAR Sample Aspect Ratio
  • Lh number of legs in the re-sampling poly-phase filter in the horizontal direction
  • N number of filter coefficients
  • Lv number of legs in the re-sampling poly-phase filter in the vertical direction
  • M filter coefficients
  • a first predetermined value indicates that a picture size in the nonbase view is the same as a picture size in the base view.
  • SAR is a ratio of width to height of a single pixel in an image in the 3D video.
  • Lh is the number of legs in the re-sampling poly-phase filter in the horizontal direction.
  • a value of ‘0’ for Lh indicates to the 3D combining module 220 that horizontal re-sampling is not required and a value of ‘1’ indicates to the 3D combining module 220 that horizontal lines are to be repeated for upsampling.
  • Lv is the number of legs in the re-sampling poly-phase filter in the vertical direction.
  • a value of ‘0’ indicates to the 3D combining module 220 that vertical re-sampling is not required and a value of ‘1’ indicates to that vertical lines are to be repeated for upsampling.
  • the information in the MPEG-2 compressed video stream includes a user-data element providing information at the picture level.
  • the values in the user-data element represent user-data type code, Length data, picture size, base view picture size, SAR, Lh, N, Lv and M.
  • the user-data type code is set to a 4-byte value signaled by a registration descriptor. A value of ‘0’ indicates to the 3D combining module 220 that picture size is the same as the base view for the nonbase view.
  • the components 210 - 220 may comprise software, hardware, or a combination of software and hardware.
  • one or more of the modules 210 - 220 comprise circuit components.
  • one or more of the modules 210 - 220 comprise software code stored on a computer readable storage medium, which is executable by a processor. It should be understood that the system 300 depicted in FIG. 3 may include additional components and that some of the components described herein may be removed and/or modified without departing from a scope of the system 300 .
  • Some or all of the operations set forth in the methods 400 - 600 may be contained as one or more computer programs stored in any desired computer readable medium and executed by a processor on a computer system.
  • Exemplary computer readable media that may be used to store software operable to implement the present invention include but are not limited to conventional computer system RAM, ROM, EPROM, EEPROM, hard disks, or other data storage devices.
  • FIG. 4 shows a method 400 of encoding a base view and a nonbase view to form a transport stream 131 .
  • the method 400 may be performed by the 3D video encoder 200 as described with respect to FIG. 2 hereinabove.
  • the 3D video encoder 200 receives video content.
  • the 3D video encoder 200 may receive the video content from, for instance, broadcast programs, Internet Protocol TV (IPTV), switched video (SDV), video on demand (VOD) or other video sources.
  • IPTV Internet Protocol TV
  • SDV switched video
  • VOD video on demand
  • the video content may include a base view and a non base view.
  • the first encoder 202 encodes the base view to form a first independently compressed video stream 132 a associated with the video content.
  • the second encoder 204 may concurrently encode the nonbase view to form a second independently compressed video stream 132 b .
  • the first and second independently compressed video streams 132 a - b may be compressed in a common compression format or alternately, in different compression formats.
  • the descriptor processor 208 generates a descriptor 209 containing information to combine the base and nonbase views to form the 3D video 222 .
  • the descriptor processor 208 may generate a 3D rendering information descriptor containing information as shown in Table 1 hereinabove.
  • the multiplexer 206 multiplexes the first and second independently compressed video streams 132 a - b to form a transport stream 131 .
  • the transport stream 131 is operable to be processed to render a 3D video using the base and nonbase views.
  • the multiplexer 206 may include the descriptor 209 received from the descriptor processor 208 in the transport stream 131 . Alternately, the multiplexer 206 may be configured to add the descriptor 209 .
  • the transport stream 131 may thereafter be output over a network to subscribers, for instance as shown in FIG. 1 .
  • FIG. 5 there is shown a method 600 of rendering a transport stream 131 into 3D video 222 .
  • the method 500 may be performed by the STB 120 c as described with respect to FIG. 3 hereinabove.
  • the STB 120 receives the transport stream 131 at the input/output module 210 .
  • the input/output module 131 of the 3D video decoder 130 is configured to receive the transport stream 131 and to send the transport stream 131 to the decoding module 214 .
  • the transport stream 131 includes first and second independently compressed video streams 132 a - b.
  • the decoding module 214 decodes the first and second independently compressed video streams 132 a - b to form a base view and a nonbase view.
  • the base view and the nonbase view are included in a same program.
  • the decoding module 214 then sends the base view and the nonbase view to the 3D combining module 220 .
  • the decoding module 214 also sends a descriptor 209 to the 3D combining module 220 .
  • the 3D combining module 220 uses the base view and the nonbase view to form 3D video 222 .
  • the descriptor 209 indicates to the 3D combining module 220 of orientations of the base and nonbase views, and of upsample rules applicable to make a resolution of the base view equal to a resolution of the nonbase view.
  • the 3D video 222 may thereafter be output to the 3D-ready TV 122 , for instance over an HDMI Interface.
  • the method 600 comprises an application of the method 500 to a transport stream 131 in which the base view and the nonbase view are compressed in a common compression format.
  • the STB 120 receives the transport stream 131 at the input/output module 210 .
  • the input/output module 131 of the 3D video decoder 130 is configured to receive the transport stream 131 and, after demultiplexing in some instances, to send the first and second independently compressed video streams 132 a - b to the decoding module 214 .
  • the decoding module 214 decodes the first and second independently compressed video streams 132 a - b to form a base view and a nonbase view.
  • the base view and the nonbase view are independently compressed and included in a same broadcast program.
  • the decoding module 214 then sends the basic stream to the 3D combining module 220 .
  • the decoding module 214 determines 3D video rendering information from a descriptor 209 , for instance as described hereinabove with respect to FIG. 2 .
  • the descriptor 209 provides values and information that may be used to render 3D video using the base view and the nonbase view.
  • the video based information includes an SEI message provided at a picture or slice level.
  • the SEI includes an orientation of the base view, an orientation of the nonbase view, and an upsample rule
  • the upsample rule is applicable to make a resolution of the base view equal to a resolution of the nonbase view.
  • the 3D combining module 220 renders the 3D video using the base view, the nonbase and the 3D rendering information included in the descriptor 209 .
  • the 3D video may thereafter be output to the 3D-ready TV 122 .
  • FIG. 7 illustrates a block diagram of a computing apparatus 700 to implement or execute one or more of the processes depicted in FIGS. 1-6 , according to an embodiment.
  • the illustration of the computing apparatus 700 is a generalized illustration and that the computing apparatus 700 may include additional components and that some of the components described may be removed and/or modified without departing from a scope of the computing apparatus 700 .
  • the computing apparatus 700 comprises the 3D video encoder 200 depicted in FIG. 2 .
  • the computing apparatus 700 comprises the STB 120 c depicted in FIG. 3 .
  • the computing apparatus 700 includes a processor 702 that may implement or execute some or all of the steps described in one or more of the processes depicted in FIGS. 4-6 . Commands and data from the processor 702 are communicated over a communication bus 704 .
  • the computing apparatus 700 also includes a main memory 706 , such as a random access memory (RAM), where the program code for the processor 702 , may be executed during runtime and a hard disk (not shown) where a copy of the program code for one or more of the processes depicted in FIGS. 4-6 may be stored.
  • the computing apparatus 700 may include non-volatile memory.
  • the computing apparatus 700 includes interfaces 714 such as coaxial cable, HDMI, USB or other network interfaces.
  • Embodiments of the present invention provide a way of rendering a transport stream into a 3D video.
  • the method and system herein disclosed allow simulcast of video components providing potential benefits over multi-view coding (MVC).
  • MVC multi-view coding
  • existing MPEG-4 or MPEG-2 decoders are capable of decoding two videos compressed using MPEG-4 or MPEG-2 and presenting the decoded stereo views for the 3D display.
  • the nonbase view that is used to create the depth perception can potentially have a lower visual quality than the base view without significantly impacting the perceptual quality of the 3D video. This can be achieved by lowering the bit rate corresponding to that the nonbase view and/or lowering the resolution of the nonbase view.
  • the nonbase can also be compressed with a different video standard than the first view, e.g. compress one of the views using MPEG-2, which can be decoded by a large number of deployed decoders, and compress the nonbase using MPEG-4. Additionally, the method and system provide cross standard and backward compatibility in systems using decoders with heterogeneous compressions standards, for instance a mixture of MPEG-2 and MPEG-4 decoders.

Abstract

A three-dimensional (3D) video encoder includes a first encoder, a second encoder, and a multiplexer. The first encoder is configured to encode a base view to form a first independently compressed video stream associated with video content. The second encoder is configured to encode a nonbase view to form a second independently compressed video stream associated with the video content. The multiplexer is configured to multiplex the first and second independently compressed video streams associated with the video content to form a transport stream. The transport stream is operable to be processed to render a 3D video using the base and nonbase views.

Description

    CLAIM OF PRIORITY
  • The present application claims the benefit of priority to U.S. Provisional Patent Application Ser. No. 61/225,792, filed on Jul. 15, 2009, entitled “Simulcast of Stereoviews For 3D TV”, the disclosure of which is hereby incorporated by reference in its entirety.
  • BACKGROUND
  • Depth perception for a three dimensional (3D) TV is provided by capturing two views, one for the left eye and other for the right eye. These two views are compressed and sent over various networks or stored on storage media. A decoder decodes the two views and sends the decoded video to the 3D TV for display. The two views are known to be either merged into a single video frame or kept separate. When kept separate, the two views are known to be compressed and distributed using various methods. Two of these methods are multi-view coding (MVC) method and simulcast method. In the MVC method, the two views are compressed such that a correlation between the two views is exploited to increase coding efficiency. In the simulcast method, the two views are known to be compressed and distributed as two independent video streams.
  • MPEG standards (such as advanced video coding (AVC) or H.264 MPEG-4 and MPEG-2 systems) provide methods and syntax to represent and compress more than one view as well as to transport this compressed data in various ways. However, the current versions of the MPEG standards do not specify how to simulcast two views so that the two views can decoded and combined appropriately to provide a 3D display. Consequently, simulcasting may not be currently done using MPEG standards for video systems and other types of systems.
  • Further, many broadcast centric standards (i.e. Society of Cable Telecommunications Engineers (SCTE) and Advanced Televisions Systems Committee (ATSC)) that are followed by a wide range of service providers and manufacturers do not allow sending two video components compressed with a same coding method within a same program or channel (i.e., two video components with the same “stream type” are not allowed in a program map table). This constraint precludes simulcasting of two video streams corresponding to the two eye views using a same compression standard in a same program. For example, the constraint precludes simulcasting two independent MPEG-2 or MPEG-4 streams (i.e. one for the left eye and one for the right eye for 3D TV) in the same program. Sending the two views in separate programs may be possible but adds complexity of requiring an additional tuner and time base synchronization between the two programs.
  • Furthermore, in many current proposals being discussed in MPEG and in the industry, both eye views are compressed using the same encoding standard. Because there are two dominant encoding standards, MPEG-2 and MPEG-4, this creates a backward compatibility problem for simulcast methods in digital TV systems that have mixture of MPEG-2 and MPEG-4 decoders. For instance, the MPEG-2 decoders will not be able to decode either view if the two views are compressed using the MPEG-4 standard.
  • SUMMARY
  • According to an embodiment, a three-dimensional (3D) video encoder includes a first encoder, a second encoder, and a multiplexer. The first encoder is configured to encode a base view to form a first independently compressed video stream associated with video content. The second encoder is configured to encode a nonbase view to form a second independently compressed video stream associated with the video content. The multiplexer is configured to multiplex the first and second independently compressed video streams associated with the video content to form a transport stream. The transport stream is operable to be processed to render a 3D video using the base and nonbase views.
  • According to another embodiment, a 3D video decoder includes an input module, a decoding module and a 3D combining module. The input module is configured to receive a transport stream. The transport stream includes first and second independently compressed video streams associated with the video content, included in a same broadcast program in the transport stream. The decoding module is configured to decode the transport stream to form base and nonbase views that are respectively from the first and second independently compressed video streams associated with the video content. The 3D combining module is configured to render a 3D video using the base and nonbase views.
  • According to another embodiment, a method of encoding a transport stream is disclosed. The method includes receiving video content. A base view is thereafter encoded to form a first independently compressed video stream associated with the video content. A nonbase view is encoded to form a second independently compressed video stream associated with the video content. The first and second independently compressed video streams associated with the video content are multiplexed to form a transport stream. The transport stream is operable to be processed to render a 3D video using the base and nonbase views.
  • According to another embodiment, a method of rendering a transport stream into 3D video is disclosed. The method includes receiving the transport stream, which includes first and second independently compressed video streams associated with the video content, included in a same broadcast program in the transport stream. The transport stream is then decoded to form base and nonbase views that are respectively formed from the first and second independently compressed video streams associated with the video content. A 3D video is thereafter rendered using the base and nonbase views.
  • Still in a further embodiment, a computer readable storage medium on which is embedded one or more computer programs implements the above-disclosed method of rendering a transport stream into 3D video.
  • Embodiments of the present invention include simulcasting independently compressed video streams in a same program for 3D TV distribution systems. The compressed video streams may have different standards, such as MPEG 2 and MPEG 4 to allow for backward compatibility. In another embodiment, the compressed video streams may have the same standards, such as both MPEG 4, to provide the highest resolution and quality 3D image in systems that have compatible decoders. Furthermore, in one embodiment, the independently compressed video streams are simulcasted in the same program by sending the streams as different types. By sending the streams in the same program rather than in two different programs, synchronization between the streams is much easier for generating the 3D video.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Features of the present invention will become apparent to those skilled in the art from the following description with reference to the figures, in which:
  • FIG. 1 illustrates a network architecture, according to an embodiment of the invention;
  • FIG. 2 illustrates a functional block diagram of multiplexing first and second independently compressed video streams associated with the video content to form a transport stream, according to an embodiment of the invention;
  • FIG. 3 illustrates a simplified block diagram of a 3D video decoder, according to an embodiment of the invention;
  • FIG. 4 illustrates a flow diagram of a method of encoding a transport stream, according to an embodiment of the invention;
  • FIG. 5 illustrates a flow diagram of a method of rendering a transport stream into 3D video, according to an embodiment of the invention, according to an embodiment of the invention;
  • FIG. 6 illustrates a flow diagram of a method of rendering a transport stream into 3D video, according to an embodiment of the invention; and
  • FIG. 7 shows a block diagram of a computer system that may be used in the 3D video decoder, according to an embodiment of the invention.
  • DETAILED DESCRIPTION
  • For simplicity and illustrative purposes, the present invention is described by referring mainly to exemplary embodiments thereof. In the following description, numerous specific details are set forth to provide a thorough understanding of the present invention. However, it will be apparent to one of ordinary skill in the art that the present invention may be practiced without limitation to these specific details. In other instances, well known methods and structures have not been described in detail to avoid unnecessarily obscuring the present invention.
  • FIG. 1 illustrates a network architecture 100 of a system in which a 3D video decoder 130 may be used, according an embodiment. As shown in FIG. 1, the network architecture 100 is illustrated as a cable television (CATV) network architecture, including a cable head-end unit 110 and a cable network 111. A number of data sources 101, 102, 103, may be communicatively coupled to the cable head-end unit 110 including, but in no way limited to, a plurality of servers 101, the Internet 102, radio signals, or television signals received via a content provider 103. The cable head-end 110 is also communicatively coupled to one or more subscribers 150 a-n through the cable network 111. It should be understood that the network architecture 100 depicted in FIG. 1 may include additional components and that some of the components described herein may be removed and/or modified without departing from a scope of the network architecture 100.
  • The cable head-end 110 is configured to output a transport stream 131. The transport stream 131 may be output to the set top boxes (STB) 120 a-n using a satellite transmission, the Internet or other network using, for instance, fixed optical fibers or coaxial cables. The transport stream 131 is configured to provide cross standard compatibility for a mixture of 3D video decoders and 2D decoders 121 that may be used by the STBs 120 a-n. Each of the subscribers 150 a-n may receive the transport stream 131 at STBs 120 a-n. The STBs 120 a-n are devices that enable TVs to receive digital television (DTV) broadcasts. In one embodiment, the STBs 120 a-n includes any device, not just conventional STBs that can perform the functions of the STBs 120 a-n described herein.
  • The STBs 120 a-n may be equipped with decoders with either 2D capability, as described with respect to the 2D decoder 121 or 3D capability. The STBs 120 a-n may be connected to either a 2D TV 123 or a 3D-ready TV 122. In order to receive and display 3D video, the STB with 3D capability may be connected to a 3D-ready TV 122.
  • It should be understood that the 3D video decoder 130 and an STB 120 c in which it is located, may be used in other networks and that the cable network architecture 100 is shown for simplicity and illustrative purposes. The 3D video decoder 130 may be used in other architectures.
  • FIG. 2 illustrates a functional block diagram of a 3D video encoder 200. As shown in FIG. 2, the 3D video encoder 200 includes a first encoder 202, a second encoder 204, a descriptor processor 208 and a multiplexer 206. It should be understood that the 3D video encoder 200 depicted in FIG. 2 may include additional components and that some of the components described herein may be removed and/or modified without departing from a scope of the 3D video encoder 200.
  • The first encoder 202 is configured to compress a base view to form a first independently compressed video stream 132 a associated with video content. A base view is an eye view for the 3D video that may be used as a full view in a two dimensional (2D) system. For instance, the base view may comprise a left eye view in a 3D video system and the full view in a 2D video system. The video content may be received from, for instance, broadcast programs, Internet Protocol TV (IPTV), switched video (SDV), video on demand (VOD) or other video sources. The second encoder is configured to compress a nonbase view to form a second independently compressed video stream 132 b associated with the video content. A nonbase view is used along with the base view to render the 3D video, as described hereinbelow, and may be dropped by 2D systems. The first and second independently compressed video streams 132 a-b associated with the video content are not compressed with reference to other components, e.g. in contrast to MVC coded video components in which the components are coded with respect to each other. Example of video compression may include conventional MPEG standards. “MPEG,” as used herein, refers to a group of standards for encoding and decoding transport streams as defined by the Moving Picture Experts Group. MPEG includes several formats for transport streams such as MPEG-2 and MPEG-4 part 10/H.264.
  • The descriptor processor 208 is configured to create and add an associated descriptor 209 to the first and second independently compressed video streams 132 a-b. The descriptor processor 208 may receive information describing the first and second independently compressed video streams 132 a-b from the first encoder 202 and the second encoder 204 respectively, and use this information in generating the descriptor 209. The descriptor 209 is configured to allow a decoder that receives the transport stream 131 to decode (including any adjustment of video resolutions) and render a 3D video from the decoded base and nonbase views and is included in both the first and second independently compressed video streams 132 a-b.
  • The multiplexer 206 is configured to receive the first and second independently compressed video streams associated with the video content 132 a-b from the first encoder 202 and the second encoder 204. The multiplexer 206 thereafter multiplexes the first and second independently compressed video streams 132 a-b along with the associated descriptor 209 to form a transport stream 131. The transport stream 131 is a video stream that includes the first and second independently compressed video streams 132 a-b. The independently compressed video streams 132 a-b may be included in a same program in the transport stream 131. A same program may correspond, for instance, to a single channel in broadcasting. The transport stream 131 may comprise Internet Protocol (IP) packets or any packetization scheme configured to transmit the transport stream 131. For instance, the transport stream 131 may comprise an MPEG stream, IP packets or any suitable medium for transporting the 3D video from a source to an end user.
  • Note that the base view in the first compression format (e.g., for video component 132 a) is a full resolution 2D compatible stream that may be received and decoded by a 2D decoder, for instance the 2D decoder 121 shown in FIG. 1 that receives the same transport stream 131. The nonbase view may be discarded by the 2D decoder 121, for instance a conventional 2D decoder. Accordingly, the transport stream 131 may be used to provide backwards compatibility in a network architecture having a mixture of STBs with conventional 2D decoders and STBs with 3D decoders the network architecture 100 with the conventional decoders while allowing 3D capability with the STB 120 c.
  • According to an embodiment, the first independently compressed video stream 132 a is configured in a first compression format, for instance MPEG-2. The second independently compressed video stream 132 b is configured in a second compression format, which differs from the first compression format, for instance MPEG-4. When the two views are compressed using different coding standards (e.g. MPEG-2 and MPEG-4) the broadcast-centric constraints allow two video components with different stream_type values within same program (i.e., the existing stream_type values for MPEG-2 video (0x02) and MPEG-4 (0x1B) can be used).
  • According to another embodiment, the first and second independently compressed video streams 132 a-b are configured in a common compression format, for instance both being MPEG-2 or both being MPEG-4. The second independently compressed video stream 132 b may be at a lower resolution than the first independently compressed video stream 132 a. The second independently compressed video stream 132 b is assigned a new stream type that is different than the first independently compressed video stream 132 a in the program map table for the transport stream 131. A program map table is used to describe each single program including elementary streams associated with the program. The stream type may be specified in the program map table. The new stream type for the nonbase view may be allocated by a transport system, for instance MPEG-2. The new stream type allocated by the MPEG-2 systems may be equal to 0x21 for MPEG-4 and equal to 0x22 for MPEG-2. Alternately, the new stream type may be determined using a next available stream type, for instance a next available ‘reserved’ stream type values in the MPEG-2 systems (International Organization for Standardization (ISO) 13818-1) stream type table.
  • The two views compressed independently may be sent (using the same compression standard such as MPEG-4) as part of the same broadcast program in the transport stream 131 if sent as two different stream types. The transport stream 131 also includes a descriptor 209 that provides information to assist in combining and rendering the two separately decoded images onto an appropriate display. The descriptor 209 provides information to assist in combining and rendering two separately decoded images from the base view and the nonbase view into a 3D image to be displayed at the 3D-ready TV 122. The descriptor 209 may specify that the independently compressed video stream is for a particular view. Thus, even though the steam type for the video stream may be reserved or some other stream type, the decoder, for instance the 3D decoder 130 in FIG. 1 hereinabove, can determine from the descriptor 209 that the stream is to be combined with another stream for the other view. The descriptor 209 may also include identification of view (e.g. left or right), upsampling rules in instances where view resolutions differ, information for combining the views (e.g. field interleaving or frame interleaving) and additional display related elements.
  • According to an embodiment, the descriptor 209 may be configured to provide information on use of single stream, combining and rendering two fields or two frames in a same video sequence and additional display related elements in instances in which the transport stream 131 is an MPEG-4 conformant stream. The single stream is used in contrast to two streams to carry the base and nonbase views.
  • According to an embodiment, the information provided in the descriptor 209 at a system level includes a 3D rendering information descriptor. The 3D rendering information descriptor includes a descriptor tag, a descriptor length and an instruction to render each picture in the 3D video. An example of the 3D rendering information descriptor in tabular form is the following:
  • TABLE 1
    Example of 3D rendering information descriptor
    No. Of
    Syntax bits Mnemonic
    3D_rendering_info_descriptor( ) {
    descriptor_tag 8 uimsbf
    descriptor_length 8 uimsbf
    two_views_in_same_stream 1 bslbf
    if (two_views_in_same_stream) {
    field_pictures 1 uimsbf
    if (field_pictures) {
    { field_parity j uimsbf
     top_field_first_in_composition k bslbf
    }
    else {frame_order l bslbf
    }
    display_specific_info m bslbf
     }
     else {
    left_or_right_view 1 uimsbf
     up_sample_horizontal n bslbf
     up_sample_vertical p bslbf
     field_or_frame_interleaving q bslbf
     display_specific_info m bslbf
    }
    }
    }
  • As shown in Table 1, j, k, l, m, n, p, and q represent constants that specify the number of bits used by each associated element in the table. The 3D rendering information descriptor provides information such as field_parity, top_field_first_in_composition, frame_order and display_specific_info including left_or_right_view, up_sample_horizontal, up_sample_vertical and field_or_frame_interleaving. For instance, by determining the descriptor constant k, the decoder determines whether the 3D video is to be rendered top field first in composition. The display specific information may include resolution of the 3D video 222. The 3D rendering information descriptor may be used by a receiver, for instance the STB 120 c, to render the received the transport stream 131 to form 3D video 222.
  • FIG. 3 illustrates a simplified block diagram of a system 300 configured to receive and display 3D content, according to an embodiment. The system 300 comprises the STB 120 c, including the 3D decoder 130, and the 3D-ready TV 122 from FIG. 1. It should be understood that one or more of the modules of the system 300 may be provided in devices other than an STB, for instance a receiver. The STB 120 c includes an input/output module 210, a demultiplexing module 212, and the 3D video encoder 130, which includes a 3D decoding module 214, and a 3D combining module 220. It should be understood that the system 300 depicted in FIG. 3 may include additional components and that some of the components described herein may be removed and/or modified without departing from a scope of the system 300.
  • The input/output module 210 is configured to receive the transport stream 131. The input/output module 210 may comprise a Universal Serial Bus (USB), an Ethernet interface, or another type of interface. The transport stream 131 received at the input/output module 210 in this instance may be a QAM modulated stream.
  • The demultiplexing module 212 may thereafter select a frequency and demodulate the frequency to obtain a multi program transport stream (MPTS). The demultiplexing module 212 is configured to demultiplex the MPTS to extract single program transport streams (SPTSs) corresponding to each of a plurality of programs which a subscriber may select. For instance, the subscriber 150 n may use the STB 120 to select a program having 3D content. The demultiplexing module 212 then demultiplexes the MPTS to form the first and second independently compressed video streams associated with the video content 132 a-b. Each of the first and second independently compressed video streams associated with the video content 132 a-b has a different codec.
  • The 3D decoding module 214 is configured to decode the first and second independently compressed video streams 132 a-b to form a base view (2D view) corresponding to a first eye view provided and a nonbase view corresponding to a second eye view. The base view and the nonbase view may be used by the STB 120 c to render a 3D video 222 and output for display at the 3D-ready TV 122.
  • The 3D combining module 220 may be configured to access the 3D rendering information descriptor in the transport stream 131 and render the 3D video 222 using the instruction therein. After decoding of the first and second independently compressed video streams associated with the video content 132 a-b by the decoding module 214, the base and nonbase views may be combined by the 3D combining module 220 to form a 3D video 222 that may be displayed on a 3D-ready TV 122 capable of displaying 3D (stereoscopic or multiple) views. The 3D combining module 220 may use a descriptor 209 included in the transport stream 131 in combining the base and nonbase views to render the 3D video 222. The descriptor 209 may include the supplemental enhancement information (SEI) to form a 3D video 222. The input/output module 210 is configured to thereafter output the 3D video 222, for instance to a 3D-ready TV.
  • The 3D combining module 220 may thereafter render pictures in the base view and the nonbase view according to information provided in the 3D rendering information descriptor. The 3D rendering information descriptor specifies operations to be performed to render the 3D video 222 from the transport stream 131. For instance, if two views are received in the same stream, the 3D combining module 220 may determine whether field pictures are received. Similarly, if field pictures are received in the base view, the 3D combining module 220 renders the 3D video 222 from the transport stream 131 according to the syntax in Table 1, hereinabove.
  • According to another embodiment, in instances in which the transport stream 131 is an MPEG-2 conformant stream, some of the rendering and combining information from the descriptor 209 may be included in the compressed video stream (such as picture user-data). This provides information on use of single stream, combining and rendering two fields or two frames in the same video sequence and additional display related elements.
  • According to another embodiment, the MPEG-4 video based rendering and combining information from the descriptor 209 includes an SEI message providing such information at a picture level or slice level in which values represent picture size, base view picture size, Sample Aspect Ratio (SAR), number of legs in the re-sampling poly-phase filter in the horizontal direction (Lh), number of filter coefficients (N) in each of those legs, number of legs in the re-sampling poly-phase filter in the vertical direction (Lv) and number of filter coefficients (M) in each of those legs. For picture size, a first predetermined value indicates that a picture size in the nonbase view is the same as a picture size in the base view. SAR is a ratio of width to height of a single pixel in an image in the 3D video. Lh is the number of legs in the re-sampling poly-phase filter in the horizontal direction. A value of ‘0’ for Lh indicates to the 3D combining module 220 that horizontal re-sampling is not required and a value of ‘1’ indicates to the 3D combining module 220 that horizontal lines are to be repeated for upsampling. A value greater than or equal to 2 indicates to the 3D combining module 220 of the coefficients (i,j), for i=0 to Lh−1, j=0 to N−1, of the poly-phase filter. Lv is the number of legs in the re-sampling poly-phase filter in the vertical direction. A value of ‘0’ indicates to the 3D combining module 220 that vertical re-sampling is not required and a value of ‘1’ indicates to that vertical lines are to be repeated for upsampling. For a value greater than or equal to 2 then the descriptor 209 provides the coefficients (i,j), for i=0 to Lv−1, j=0 to M−1, of the poly-phase filter.
  • According to another embodiment, the information in the MPEG-2 compressed video stream includes a user-data element providing information at the picture level. The values in the user-data element represent user-data type code, Length data, picture size, base view picture size, SAR, Lh, N, Lv and M. The user-data type code is set to a 4-byte value signaled by a registration descriptor. A value of ‘0’ indicates to the 3D combining module 220 that picture size is the same as the base view for the nonbase view.
  • The components 210-220 may comprise software, hardware, or a combination of software and hardware. Thus, in one embodiment, one or more of the modules 210-220 comprise circuit components. In another embodiment, one or more of the modules 210-220 comprise software code stored on a computer readable storage medium, which is executable by a processor. It should be understood that the system 300 depicted in FIG. 3 may include additional components and that some of the components described herein may be removed and/or modified without departing from a scope of the system 300.
  • Examples of methods in which the 3D video encoder 200 and the decoder 300 may be employed for simulcast of stereoviews in 3D TV are now described with respect to the following flow diagrams of the methods 400-600 depicted in FIGS. 4-6. It should be apparent to those of ordinary skill in the art that the methods 400-600 represent generalized illustrations and that other steps may be added or existing steps may be removed, modified or rearranged without departing from the scopes of the methods 400-600. In addition, the methods 400-600 are described with respect to the 3D video encoder 200 and the decoder 300 by way of example and not limitation, and the methods 400-600 may be used in other systems.
  • Some or all of the operations set forth in the methods 400-600 may be contained as one or more computer programs stored in any desired computer readable medium and executed by a processor on a computer system. Exemplary computer readable media that may be used to store software operable to implement the present invention include but are not limited to conventional computer system RAM, ROM, EPROM, EEPROM, hard disks, or other data storage devices.
  • FIG. 4, shows a method 400 of encoding a base view and a nonbase view to form a transport stream 131. The method 400 may be performed by the 3D video encoder 200 as described with respect to FIG. 2 hereinabove.
  • At step 402, as shown in FIG. 4, the 3D video encoder 200 receives video content. The 3D video encoder 200 may receive the video content from, for instance, broadcast programs, Internet Protocol TV (IPTV), switched video (SDV), video on demand (VOD) or other video sources. The video content may include a base view and a non base view.
  • At step 404, the first encoder 202 encodes the base view to form a first independently compressed video stream 132 a associated with the video content. At step 406, the second encoder 204 may concurrently encode the nonbase view to form a second independently compressed video stream 132 b. The first and second independently compressed video streams 132 a-b may be compressed in a common compression format or alternately, in different compression formats.
  • At step 408, the descriptor processor 208 generates a descriptor 209 containing information to combine the base and nonbase views to form the 3D video 222. For instance, the descriptor processor 208, as shown with respect to FIG. 2 hereinabove, may generate a 3D rendering information descriptor containing information as shown in Table 1 hereinabove.
  • At step 410, the multiplexer 206 multiplexes the first and second independently compressed video streams 132 a-b to form a transport stream 131. The transport stream 131 is operable to be processed to render a 3D video using the base and nonbase views. The multiplexer 206 may include the descriptor 209 received from the descriptor processor 208 in the transport stream 131. Alternately, the multiplexer 206 may be configured to add the descriptor 209. The transport stream 131 may thereafter be output over a network to subscribers, for instance as shown in FIG. 1.
  • Turning now to FIG. 5, there is shown a method 600 of rendering a transport stream 131 into 3D video 222. The method 500 may be performed by the STB 120 c as described with respect to FIG. 3 hereinabove.
  • At step 502, as shown in FIG. 5, the STB 120 receives the transport stream 131 at the input/output module 210. The input/output module 131 of the 3D video decoder 130 is configured to receive the transport stream 131 and to send the transport stream 131 to the decoding module 214. The transport stream 131 includes first and second independently compressed video streams 132 a-b.
  • At step 504, the decoding module 214 decodes the first and second independently compressed video streams 132 a-b to form a base view and a nonbase view. The base view and the nonbase view are included in a same program. The decoding module 214 then sends the base view and the nonbase view to the 3D combining module 220. The decoding module 214 also sends a descriptor 209 to the 3D combining module 220.
  • At step 506, the 3D combining module 220 uses the base view and the nonbase view to form 3D video 222. The descriptor 209 indicates to the 3D combining module 220 of orientations of the base and nonbase views, and of upsample rules applicable to make a resolution of the base view equal to a resolution of the nonbase view. The 3D video 222 may thereafter be output to the 3D-ready TV 122, for instance over an HDMI Interface.
  • Turning now to FIG. 6, there is shown a method 600 of rendering a transport stream 131 into 3D video 222. The method 600 comprises an application of the method 500 to a transport stream 131 in which the base view and the nonbase view are compressed in a common compression format.
  • At step 602, as shown in FIG. 6, the STB 120 receives the transport stream 131 at the input/output module 210. The input/output module 131 of the 3D video decoder 130 is configured to receive the transport stream 131 and, after demultiplexing in some instances, to send the first and second independently compressed video streams 132 a-b to the decoding module 214.
  • At step 604, the decoding module 214 decodes the first and second independently compressed video streams 132 a-b to form a base view and a nonbase view. The base view and the nonbase view are independently compressed and included in a same broadcast program. The decoding module 214 then sends the basic stream to the 3D combining module 220.
  • At step 606, the decoding module 214 determines 3D video rendering information from a descriptor 209, for instance as described hereinabove with respect to FIG. 2. The descriptor 209 provides values and information that may be used to render 3D video using the base view and the nonbase view.
  • According to an embodiment, the video based information includes an SEI message provided at a picture or slice level. The SEI includes an orientation of the base view, an orientation of the nonbase view, and an upsample rule The upsample rule is applicable to make a resolution of the base view equal to a resolution of the nonbase view.
  • At step 608, the 3D combining module 220 renders the 3D video using the base view, the nonbase and the 3D rendering information included in the descriptor 209. The 3D video may thereafter be output to the 3D-ready TV 122.
  • Although described specifically throughout the entirety of the instant disclosure, representative embodiments of the present invention have utility over a wide range of applications, and the above discussion is not intended and should not be construed to be limiting, but is offered as an illustrative discussion of aspects of the invention.
  • FIG. 7 illustrates a block diagram of a computing apparatus 700 to implement or execute one or more of the processes depicted in FIGS. 1-6, according to an embodiment. It should be understood that the illustration of the computing apparatus 700 is a generalized illustration and that the computing apparatus 700 may include additional components and that some of the components described may be removed and/or modified without departing from a scope of the computing apparatus 700. According to an embodiment, the computing apparatus 700 comprises the 3D video encoder 200 depicted in FIG. 2. According to another embodiment, the computing apparatus 700 comprises the STB 120 c depicted in FIG. 3.
  • The computing apparatus 700 includes a processor 702 that may implement or execute some or all of the steps described in one or more of the processes depicted in FIGS. 4-6. Commands and data from the processor 702 are communicated over a communication bus 704. The computing apparatus 700 also includes a main memory 706, such as a random access memory (RAM), where the program code for the processor 702, may be executed during runtime and a hard disk (not shown) where a copy of the program code for one or more of the processes depicted in FIGS. 4-6 may be stored. Alternately, the computing apparatus 700 may include non-volatile memory. In addition, the computing apparatus 700 includes interfaces 714 such as coaxial cable, HDMI, USB or other network interfaces.
  • Embodiments of the present invention provide a way of rendering a transport stream into a 3D video. The method and system herein disclosed allow simulcast of video components providing potential benefits over multi-view coding (MVC). In MVC, the resolution of both views in a 3D transmission must be kept the same. Existing MPEG-4 or MPEG-2 decoders are capable of decoding two videos compressed using MPEG-4 or MPEG-2 and presenting the decoded stereo views for the 3D display. The nonbase view that is used to create the depth perception can potentially have a lower visual quality than the base view without significantly impacting the perceptual quality of the 3D video. This can be achieved by lowering the bit rate corresponding to that the nonbase view and/or lowering the resolution of the nonbase view. The nonbase can also be compressed with a different video standard than the first view, e.g. compress one of the views using MPEG-2, which can be decoded by a large number of deployed decoders, and compress the nonbase using MPEG-4. Additionally, the method and system provide cross standard and backward compatibility in systems using decoders with heterogeneous compressions standards, for instance a mixture of MPEG-2 and MPEG-4 decoders.
  • What has been described and illustrated herein are embodiments of the invention along with some of their variations. The terms, descriptions and figures used herein are set forth by way of illustration only and are not meant as limitations. Those skilled in the art will recognize that many variations are possible within the spirit and scope of the invention, wherein the invention is intended to be defined by the following claims—and their equivalents—in which all terms are meant in their broadest reasonable sense unless otherwise indicated.

Claims (23)

What is claimed is:
1. A three dimensional (3D) video encoder comprising:
a first encoder configured to encode a base view to form a first independently compressed video stream associated with video content;
a second encoder configured to encode a nonbase view to form a second independently compressed video stream associated with the video content; and
a multiplexer configured to multiplex the first and second independently compressed video streams associated with the video content to form a transport stream, wherein the transport stream is operable to be processed to render a 3D video using the base and nonbase views.
2. The 3D video encoder of claim 1, further comprising:
a descriptor processor configured to generate a descriptor associated with the first and second independently compressed video streams, wherein the descriptor contains information for use in combining the base and nonbase views to form the 3D video and wherein the multiplexer is configured to include the descriptor in the transport stream.
3. The 3D video encoder of claim 2, wherein the first and second independently compressed video streams associated with the video content are encoded under a common compression format and assigned different stream types, and wherein the common compression format is one of an MPEG-4 and an MPEG-2 format, and wherein the descriptor contains information for use in rendering at least two fields in a common video sequence, wherein the at least two fields are rendered from both the first and second independently compressed video streams.
4. The 3D video encoder of claim 2, wherein the information contained in the descriptor includes orientations of the base and nonbase views, and an upsample rule that is applicable to make a resolution of the base view equal to a resolution of the nonbase view.
5. The 3D video encoder of claim 2, wherein the first independently compressed video stream is operable to be processed by a 2D decoder to form a full resolution 2D compatible stream, and wherein the 2D decoder is operable to drop the second compressed video stream.
6. The 3D video encoder of claim 2, wherein the transport stream includes a supplemental enhancement information (SEI) message providing information at a picture level or slice level in which values represent one or more of:
picture size, wherein a first predetermined value indicates that a picture size in the nonbase view is the same as a picture size in the base view;
base view picture size;
Sample Aspect Ratio (SAR);
Lh, wherein Lh is a number of legs in a re-sampling poly-phase filter in a horizontal direction, and wherein,
a value of ‘0’ indicates that horizontal re-sampling is not required,
a value of ‘1’ indicates that horizontal lines are to be repeated for upsampling,
a value greater than or equal to 2 indicates that coefficients (i,j) of the re-sampling poly-phase filter are to set to, for i=0 to Lh−1, j=0 to N−1, wherein N is a number of coefficients in each of the poly-phase legs; and
Lv, wherein Lv is a number of legs in the re-sampling poly-phase filter in a vertical direction, and wherein
a value of ‘0’ indicates that vertical re-sampling is not required,
a value of ‘1’ indicates that vertical lines are to be repeated for upsampling,
a value greater than or equal to 2 indicates that coefficients (i,j) of the re-sampling poly-phase filter are to be set, for i=0 to Lv−1, j=0 to M−1, wherein M is a number of coefficients in each of the poly-phase legs.
7. The 3D video encoder of claim 2, wherein the transport stream includes a user-data element providing information at a picture level in which values represent one or more of:
user-data type code set to a 4-byte value signaled by a registration descriptor;
Length data;
picture size, wherein a value of ‘0’ indicates the picture size is the same for the base and nonbase views;
base view picture size;
Sample Aspect Ratio (SAR);
Lh, wherein Lh is a number of legs in a re-sampling poly-phase filter in a horizontal direction, and wherein,
a value of ‘0’ indicates that horizontal re-sampling is not required,
a value of ‘1’ indicates that horizontal lines are to be repeated for upsampling,
a value greater than or equal to 2 indicates that coefficients (i,j) of the re-sampling poly-phase filter are set to, for i=0 to Lh−1, j=0 to N−1, wherein N is a number of coefficients in each of the poly-phase legs; and
Lv, wherein Lv is a number of legs in the re-sampling poly-phase filter in a vertical direction and wherein,
a value of ‘0’ indicates that vertical re-sampling is not required,
a value of ‘1’ indicates that vertical lines are to be repeated for upsampling,
a value greater than or equal to 2 indicates that coefficients (i,j) of the re-sampling poly-phase filter are set to, for i=0 to Lv−1, j=0 to M−1, wherein M is the number of coefficients in each of the poly-phase legs.
8. A 3D video decoder comprising:
an input module configured to receive a transport stream including first and second independently compressed streams associated with video content;
a decoding module configured to decode the transport stream to form base and nonbase views respectively from the first and second independently compressed video streams; and
a 3D combining module configured to render a 3D video using the base and nonbase views.
9. The 3D video decoder of claim 8, wherein the transport stream includes a descriptor providing information to combine the base and nonbase views to form the 3D video; and
the first and second independently compressed video streams associated with the video content are separately coded in a common compression format and assigned different stream types.
10. The 3D video decoder of claim 8, the transport stream includes a descriptor providing information to combine the base and nonbase views to form the 3D video; and
the first and second independently compressed video streams associated with the video content are respectively coded in first and second compression formats that are different from each other.
11. A receiver comprising:
a processor configured to receive a transport stream including first and second independently compressed video streams associated with video content, wherein the transport stream includes a descriptor operable by the processor to render a 3D video, the descriptor having syntax including
a descriptor tag indentifying the descriptor;
a descriptor length; and
instructions for rendering the 3D video, the instructions including
if two views are in the transport stream, then determine if field pictures are received in the transport stream, and if field pictures are received then determine if field parity exists, and if field parity exists then use top field first in composition, if field pictures are not received, use frame order and use display specific information; and
if two views are not in the transport stream, then up sample horizontally, up sample vertically, use field or frame interleaving, and use display specific information for a left eye view or a right eye view in the transport stream.
12. A method of encoding a transport stream, the method comprising:
receiving video content;
encoding a base view to form a first independently compressed video stream associated with the video content;
encoding a nonbase view to form a second independently compressed video stream associated with the video content; and
multiplexing the first and second independently compressed video streams associated with the video content to form a transport stream, wherein the transport stream is operable to be processed to render a 3D video using the base and nonbase views.
13. The method of claim 12, further comprising:
generating a descriptor, wherein the descriptor contains information for use in combining the base and nonbase views to form the 3D video; and
wherein multiplexing the first and second independently compressed video streams associated with the video content to form the transport stream comprises including a descriptor in the transport stream for containing information to combine the base and nonbase views to form the 3D video.
14. The method of claim 13, wherein encoding the base view to form the first independently compressed video stream associated with the video content, and encoding the nonbase view to form the second independently compressed video stream associated with the video content further comprises:
separately coding the first and second independently compressed video streams associated with the video content in a common compression format and assigning different stream types to the first and second independently compressed video streams.
15. The method of claim 13, wherein encoding the base view to form the first independently compressed video stream associated with the video content, and encoding the nonbase view to form the second independently compressed video stream associated with the video content further comprises:
respectively coding the first and second independently compressed video streams associated with the video content in first and second compression formats that are different from each other.
16. The method of claim 13, wherein including the descriptor in the transport stream further comprises:
including an SEI message providing user-data information at a picture level or slice level in which values represent
picture size, wherein a first predetermined value indicates that a picture size in the nonbase view is the same as a picture size in the base view;
base view picture size;
Sample Aspect Ratio (SAR);
Lh, wherein Lh is a number of legs in a re-sampling poly-phase filter in a horizontal direction, and wherein
a value of ‘0’ indicates that horizontal re-sampling is not required,
a value of ‘1’ indicates that horizontal lines are to be repeated for upsampling,
a value greater than or equal to 2 indicates that coefficients (i,j) of the re-sampling poly-phase filter are to set to, for i=0 to Lh−1, j=0 to N−1, wherein N is a number of coefficients in each of the poly-phase legs; and
Lv, wherein Lv is a number of legs in the re-sampling poly-phase filter in a vertical direction, and wherein
a value of ‘0’ indicates that vertical re-sampling is not required,
a value of ‘1’ indicates that vertical lines are to be repeated for upsampling,
a value greater than or equal to 2 indicates that coefficients (i,j) of the re-sampling poly-phase filter are to be set, for i=0 to Lv−1, j=0 to M−1, wherein M is a number of coefficients in each of the poly-phase legs.
17. The method of claim 13, wherein including the descriptor in the transport stream further comprises:
including a user-data element providing information at a picture level in which values represent one or more of:
user-data type code set to a 4-byte value signaled by a registration descriptor;
Length data;
picture size, wherein a value of ‘0’ indicates the picture size is the same for the base and nonbase views;
base view picture size;
Sample Aspect Ratio (SAR);
Lh, wherein Lh is a number of legs in a re-sampling poly-phase filter in a horizontal direction, and wherein,
a value of ‘0’ indicates that horizontal re-sampling is not required,
a value of ‘1’ indicates that horizontal lines are to be repeated for upsampling,
a value greater than or equal to 2 indicates that coefficients (i,j) of the re-sampling poly-phase filter are set to, for i=0 to Lh−1, j=0 to N−1, wherein N is a number of coefficients in each of the poly-phase legs; and
Lv, wherein Lv is the number of legs in the re-sampling poly-phase filter in the vertical direction and wherein
a value of ‘0’ indicates that vertical re-sampling is not required,
a value of ‘1’ indicates that vertical lines are to be repeated for upsampling,
a value greater than or equal to 2 indicates that coefficients (i,j) of the re-sampling poly-phase filter are set to, for i=0 to Lv−1, j=0 to M−1, wherein M is the number of coefficients in each of the poly-phase legs.
18. The method of claim 13, wherein multiplexing the first and second independently compressed video streams associated with the video content to form the transport stream further comprises:
including a descriptor in the transport stream for providing information to combine the base and nonbase views to form the 3D video; and
wherein the first and second independently compressed video streams associated with the video content are separately coded in first and second compression formats that are different from each other.
19. A method of rendering a transport stream into 3D video, the method comprising:
receiving the transport stream including first and second independently compressed video streams associated with video content;
decoding the transport stream to form base and nonbase views that are respectively formed from the first and second independently compressed video streams associated with the video content; and
rendering a 3D video using the base and nonbase views.
20. The method of claim 19, wherein receiving the transport stream further comprises:
receiving a descriptor in the transport stream for providing information to combine the base and nonbase views to form the 3D video; and
receiving the first and second independently compressed video streams associated with the video content separately coded in a common compression format, the first and second independently compressed video streams associated with the video content being assigned different stream types.
21. The method of claim 19, wherein receiving the transport stream further comprises:
receiving a descriptor in the transport stream for providing information to combine the base and nonbase views to form the 3D video; and
receiving the first and second independently compressed video streams associated with the video content separately coded in first and second compression formats that are different from each other.
22. A computer readable storage medium storing at least one computer program that when executed performs a method of rendering a transport stream into 3D video, the method comprising:
receiving the transport stream including first and second independently compressed video streams associated with the video content that are included in a same broadcast program in the transport stream;
decoding the transport stream to form base and nonbase views that are respectively formed from the first and second independently compressed video streams associated with the video content; and
rendering a 3D video using the base and nonbase views.
23. A computer readable storage medium storing at least one computer program that when executed performs a method of encoding a transport stream, the method comprising:
encoding a base view to form a first independently compressed video stream associated with video content;
encoding a nonbase view to form a second independently compressed video stream associated with the video content; and
multiplexing the first and second independently compressed video streams associated with the video content to form a transport stream, wherein the transport stream is operable to be processed to render a 3D video using the base and nonbase views.
US12/835,103 2009-07-15 2010-07-13 Simulcast of stereoviews for 3D TV Active 2032-09-26 US9036700B2 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US12/835,103 US9036700B2 (en) 2009-07-15 2010-07-13 Simulcast of stereoviews for 3D TV
PCT/US2010/042072 WO2011008917A1 (en) 2009-07-15 2010-07-15 Simulcast of stereoviews for 3d tv
CN201080031847.6A CN102474635B (en) 2009-07-15 2010-07-15 Simulcast of stereoviews for 3D TV
KR1020127003916A KR101342294B1 (en) 2009-07-15 2010-07-15 Simulcast of stereoviews for 3d tv
EP10800518.2A EP2454886B1 (en) 2009-07-15 2010-07-15 Simulcast of stereoviews for 3d tv

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US22579209P 2009-07-15 2009-07-15
US12/835,103 US9036700B2 (en) 2009-07-15 2010-07-13 Simulcast of stereoviews for 3D TV

Publications (2)

Publication Number Publication Date
US20110012992A1 true US20110012992A1 (en) 2011-01-20
US9036700B2 US9036700B2 (en) 2015-05-19

Family

ID=43449782

Family Applications (1)

Application Number Title Priority Date Filing Date
US12/835,103 Active 2032-09-26 US9036700B2 (en) 2009-07-15 2010-07-13 Simulcast of stereoviews for 3D TV

Country Status (5)

Country Link
US (1) US9036700B2 (en)
EP (1) EP2454886B1 (en)
KR (1) KR101342294B1 (en)
CN (1) CN102474635B (en)
WO (1) WO2011008917A1 (en)

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120062698A1 (en) * 2010-09-08 2012-03-15 Electronics And Telecommunications Research Institute Apparatus and method for transmitting/receiving data in communication system
US20120075418A1 (en) * 2010-09-27 2012-03-29 Samsung Electronics Co., Ltd. Video processing apparatus, content providing server, and control method thereof
US20120206568A1 (en) * 2011-02-10 2012-08-16 Google Inc. Computing device having multiple image capture devices and image modes
US20120320036A1 (en) * 2011-06-17 2012-12-20 Lg Display Co., Ltd. Stereoscopic Image Display Device and Driving Method Thereof
US8438502B2 (en) 2010-08-25 2013-05-07 At&T Intellectual Property I, L.P. Apparatus for controlling three-dimensional images
US20130208090A1 (en) * 2010-10-26 2013-08-15 Korean Broadcasting System Hierarchical broadcasting system and method for 3d broadcasting
US20130250051A1 (en) * 2010-12-13 2013-09-26 Electronics And Telecommunications Research Institute Signaling method for a stereoscopic video service and apparatus using the method
US8587635B2 (en) 2011-07-15 2013-11-19 At&T Intellectual Property I, L.P. Apparatus and method for providing media services with telepresence
US8593574B2 (en) 2010-06-30 2013-11-26 At&T Intellectual Property I, L.P. Apparatus and method for providing dimensional media content based on detected display capability
US20140015927A1 (en) * 2010-01-28 2014-01-16 Samsung Electronics Co., Ltd. Method and apparatus for transmitting digital broadcasting stream using linking information about multi-view video stream, and method and apparatus for receiving the same
US8640182B2 (en) 2010-06-30 2014-01-28 At&T Intellectual Property I, L.P. Method for detecting a viewing apparatus
US20140125780A1 (en) * 2010-12-08 2014-05-08 L G Electronics Inc. Device and method for receiving digital broadcast signal
US20140269937A1 (en) * 2013-03-14 2014-09-18 Harris Corporation Systems and methods for multiple stream encoded digital video
US20140340490A1 (en) * 2013-05-15 2014-11-20 Paul Duffy Portable simulated 3d projection apparatus
US8918831B2 (en) 2010-07-06 2014-12-23 At&T Intellectual Property I, Lp Method and apparatus for managing a presentation of media content
US8947497B2 (en) 2011-06-24 2015-02-03 At&T Intellectual Property I, Lp Apparatus and method for managing telepresence sessions
US8947511B2 (en) 2010-10-01 2015-02-03 At&T Intellectual Property I, L.P. Apparatus and method for presenting three-dimensional media content
US20150089564A1 (en) * 2012-04-23 2015-03-26 Lg Electronics Inc. Signal processing device and method for 3d service
US8994716B2 (en) 2010-08-02 2015-03-31 At&T Intellectual Property I, Lp Apparatus and method for providing media content
US9030536B2 (en) 2010-06-04 2015-05-12 At&T Intellectual Property I, Lp Apparatus and method for presenting media content
US9030522B2 (en) 2011-06-24 2015-05-12 At&T Intellectual Property I, Lp Apparatus and method for providing media content
US9032470B2 (en) 2010-07-20 2015-05-12 At&T Intellectual Property I, Lp Apparatus for adapting a presentation of media content according to a position of a viewing apparatus
US9049426B2 (en) 2010-07-07 2015-06-02 At&T Intellectual Property I, Lp Apparatus and method for distributing three dimensional media content
US20150229900A1 (en) * 2009-04-07 2015-08-13 Lg Electronics, Inc. Broadcast transmitter, broadcast receiver and 3d video data processing method thereof
US9232274B2 (en) 2010-07-20 2016-01-05 At&T Intellectual Property I, L.P. Apparatus for adapting a presentation of media content to a requesting device
US9445046B2 (en) 2011-06-24 2016-09-13 At&T Intellectual Property I, L.P. Apparatus and method for presenting media content with telepresence
US9560406B2 (en) 2010-07-20 2017-01-31 At&T Intellectual Property I, L.P. Method and apparatus for adapting a presentation of media content
US9602766B2 (en) 2011-06-24 2017-03-21 At&T Intellectual Property I, L.P. Apparatus and method for presenting three dimensional objects with telepresence
US9787974B2 (en) * 2010-06-30 2017-10-10 At&T Intellectual Property I, L.P. Method and apparatus for delivering media content
US10368052B2 (en) * 2011-05-24 2019-07-30 Comcast Cable Communications, Llc Dynamic distribution of three-dimensional content
US10448083B2 (en) 2010-04-06 2019-10-15 Comcast Cable Communications, Llc Streaming and rendering of 3-dimensional video
US11711592B2 (en) * 2010-04-06 2023-07-25 Comcast Cable Communications, Llc Distribution of multiple signals of video content independently over a network

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10152828B2 (en) * 2015-09-30 2018-12-11 Umap AV Corp. Generating scene reconstructions from images
FI20165115A (en) * 2016-02-17 2017-08-18 Nokia Technologies Oy Hardware, method and computer program for video encoding and decoding
WO2023025980A1 (en) * 2021-08-26 2023-03-02 Nokia Technologies Oy An apparatus, a method and a computer program for volumetric video

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4802005A (en) * 1986-07-30 1989-01-31 Sony Corporation High efficiency encoding system
US20020009137A1 (en) * 2000-02-01 2002-01-24 Nelson John E. Three-dimensional video broadcasting system
US20030067637A1 (en) * 2000-05-15 2003-04-10 Nokia Corporation Video coding
US6704042B2 (en) * 1998-12-10 2004-03-09 Canon Kabushiki Kaisha Video processing apparatus, control method therefor, and storage medium
US20050220192A1 (en) * 2002-08-27 2005-10-06 Hsiang-Chun Huang Architecture and method for fine granularity scalable video coding
US20060221178A1 (en) * 2003-04-17 2006-10-05 Kug-Jin Yun System and method for internet broadcasting of mpeg-4-based stereoscopic video
US20070041444A1 (en) * 2004-02-27 2007-02-22 Gutierrez Novelo Manuel R Stereoscopic 3D-video image digital decoding system and method
US20070211796A1 (en) * 2006-03-09 2007-09-13 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding multi-view video to provide uniform picture quality
US20080317124A1 (en) * 2007-06-25 2008-12-25 Sukhee Cho Multi-view video coding system, decoding system, bitstream extraction system for decoding base view and supporting view random access
US20090288125A1 (en) * 2005-07-15 2009-11-19 Yoshihiro Morioka Packet transmitting apparatus
US20100110163A1 (en) * 2007-09-24 2010-05-06 Koninklijke Philips Electronics N.V. Method and system for encoding a video data signal, encoded video data signal, method and sytem for decoding a video data signal
US20110001792A1 (en) * 2008-03-04 2011-01-06 Purvin Bibhas Pandit Virtual reference view
US20110286530A1 (en) * 2009-01-26 2011-11-24 Dong Tian Frame packing for video coding

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6765568B2 (en) 2000-06-12 2004-07-20 Vrex, Inc. Electronic stereoscopic media delivery system
KR100658222B1 (en) * 2004-08-09 2006-12-15 한국전자통신연구원 3 Dimension Digital Multimedia Broadcasting System

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4802005A (en) * 1986-07-30 1989-01-31 Sony Corporation High efficiency encoding system
US6704042B2 (en) * 1998-12-10 2004-03-09 Canon Kabushiki Kaisha Video processing apparatus, control method therefor, and storage medium
US20020009137A1 (en) * 2000-02-01 2002-01-24 Nelson John E. Three-dimensional video broadcasting system
US20030067637A1 (en) * 2000-05-15 2003-04-10 Nokia Corporation Video coding
US20050220192A1 (en) * 2002-08-27 2005-10-06 Hsiang-Chun Huang Architecture and method for fine granularity scalable video coding
US20060221178A1 (en) * 2003-04-17 2006-10-05 Kug-Jin Yun System and method for internet broadcasting of mpeg-4-based stereoscopic video
US20070041444A1 (en) * 2004-02-27 2007-02-22 Gutierrez Novelo Manuel R Stereoscopic 3D-video image digital decoding system and method
US20090288125A1 (en) * 2005-07-15 2009-11-19 Yoshihiro Morioka Packet transmitting apparatus
US7929560B2 (en) * 2005-07-15 2011-04-19 Panasonic Corporation Packet transmitting apparatus
US20070211796A1 (en) * 2006-03-09 2007-09-13 Samsung Electronics Co., Ltd. Method and apparatus for encoding and decoding multi-view video to provide uniform picture quality
US20080317124A1 (en) * 2007-06-25 2008-12-25 Sukhee Cho Multi-view video coding system, decoding system, bitstream extraction system for decoding base view and supporting view random access
US20100110163A1 (en) * 2007-09-24 2010-05-06 Koninklijke Philips Electronics N.V. Method and system for encoding a video data signal, encoded video data signal, method and sytem for decoding a video data signal
US20110001792A1 (en) * 2008-03-04 2011-01-06 Purvin Bibhas Pandit Virtual reference view
US20110286530A1 (en) * 2009-01-26 2011-11-24 Dong Tian Frame packing for video coding

Cited By (68)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9762885B2 (en) * 2009-04-07 2017-09-12 Lg Electronics Inc. Broadcast transmitter, broadcast receiver and 3D video data processing method thereof
US9756311B2 (en) 2009-04-07 2017-09-05 Lg Electronics Inc. Broadcast transmitter, broadcast receiver and 3D video data processing method thereof
US10129525B2 (en) 2009-04-07 2018-11-13 Lg Electronics Inc. Broadcast transmitter, broadcast receiver and 3D video data processing method thereof
US20150229900A1 (en) * 2009-04-07 2015-08-13 Lg Electronics, Inc. Broadcast transmitter, broadcast receiver and 3d video data processing method thereof
US20140015927A1 (en) * 2010-01-28 2014-01-16 Samsung Electronics Co., Ltd. Method and apparatus for transmitting digital broadcasting stream using linking information about multi-view video stream, and method and apparatus for receiving the same
US10448083B2 (en) 2010-04-06 2019-10-15 Comcast Cable Communications, Llc Streaming and rendering of 3-dimensional video
US11368741B2 (en) 2010-04-06 2022-06-21 Comcast Cable Communications, Llc Streaming and rendering of multidimensional video using a plurality of data streams
US11711592B2 (en) * 2010-04-06 2023-07-25 Comcast Cable Communications, Llc Distribution of multiple signals of video content independently over a network
US9380294B2 (en) 2010-06-04 2016-06-28 At&T Intellectual Property I, Lp Apparatus and method for presenting media content
US9030536B2 (en) 2010-06-04 2015-05-12 At&T Intellectual Property I, Lp Apparatus and method for presenting media content
US9774845B2 (en) 2010-06-04 2017-09-26 At&T Intellectual Property I, L.P. Apparatus and method for presenting media content
US10567742B2 (en) 2010-06-04 2020-02-18 At&T Intellectual Property I, L.P. Apparatus and method for presenting media content
US8593574B2 (en) 2010-06-30 2013-11-26 At&T Intellectual Property I, L.P. Apparatus and method for providing dimensional media content based on detected display capability
US8640182B2 (en) 2010-06-30 2014-01-28 At&T Intellectual Property I, L.P. Method for detecting a viewing apparatus
US9787974B2 (en) * 2010-06-30 2017-10-10 At&T Intellectual Property I, L.P. Method and apparatus for delivering media content
US9781469B2 (en) 2010-07-06 2017-10-03 At&T Intellectual Property I, Lp Method and apparatus for managing a presentation of media content
US8918831B2 (en) 2010-07-06 2014-12-23 At&T Intellectual Property I, Lp Method and apparatus for managing a presentation of media content
US9049426B2 (en) 2010-07-07 2015-06-02 At&T Intellectual Property I, Lp Apparatus and method for distributing three dimensional media content
US11290701B2 (en) 2010-07-07 2022-03-29 At&T Intellectual Property I, L.P. Apparatus and method for distributing three dimensional media content
US10237533B2 (en) 2010-07-07 2019-03-19 At&T Intellectual Property I, L.P. Apparatus and method for distributing three dimensional media content
US9560406B2 (en) 2010-07-20 2017-01-31 At&T Intellectual Property I, L.P. Method and apparatus for adapting a presentation of media content
US10602233B2 (en) 2010-07-20 2020-03-24 At&T Intellectual Property I, L.P. Apparatus for adapting a presentation of media content to a requesting device
US10489883B2 (en) 2010-07-20 2019-11-26 At&T Intellectual Property I, L.P. Apparatus for adapting a presentation of media content according to a position of a viewing apparatus
US9032470B2 (en) 2010-07-20 2015-05-12 At&T Intellectual Property I, Lp Apparatus for adapting a presentation of media content according to a position of a viewing apparatus
US9232274B2 (en) 2010-07-20 2016-01-05 At&T Intellectual Property I, L.P. Apparatus for adapting a presentation of media content to a requesting device
US10070196B2 (en) 2010-07-20 2018-09-04 At&T Intellectual Property I, L.P. Apparatus for adapting a presentation of media content to a requesting device
US9668004B2 (en) 2010-07-20 2017-05-30 At&T Intellectual Property I, L.P. Apparatus for adapting a presentation of media content to a requesting device
US9830680B2 (en) 2010-07-20 2017-11-28 At&T Intellectual Property I, L.P. Apparatus for adapting a presentation of media content according to a position of a viewing apparatus
US9247228B2 (en) 2010-08-02 2016-01-26 At&T Intellectual Property I, Lp Apparatus and method for providing media content
US8994716B2 (en) 2010-08-02 2015-03-31 At&T Intellectual Property I, Lp Apparatus and method for providing media content
US9086778B2 (en) 2010-08-25 2015-07-21 At&T Intellectual Property I, Lp Apparatus for controlling three-dimensional images
US8438502B2 (en) 2010-08-25 2013-05-07 At&T Intellectual Property I, L.P. Apparatus for controlling three-dimensional images
US9352231B2 (en) 2010-08-25 2016-05-31 At&T Intellectual Property I, Lp Apparatus for controlling three-dimensional images
US9700794B2 (en) 2010-08-25 2017-07-11 At&T Intellectual Property I, L.P. Apparatus for controlling three-dimensional images
US20120062698A1 (en) * 2010-09-08 2012-03-15 Electronics And Telecommunications Research Institute Apparatus and method for transmitting/receiving data in communication system
US20120075418A1 (en) * 2010-09-27 2012-03-29 Samsung Electronics Co., Ltd. Video processing apparatus, content providing server, and control method thereof
US8810630B2 (en) * 2010-09-27 2014-08-19 Samsung Electronics Co., Ltd. Video processing apparatus, content providing server, and control method thereof
US8947511B2 (en) 2010-10-01 2015-02-03 At&T Intellectual Property I, L.P. Apparatus and method for presenting three-dimensional media content
US20130208090A1 (en) * 2010-10-26 2013-08-15 Korean Broadcasting System Hierarchical broadcasting system and method for 3d broadcasting
US9628780B2 (en) * 2010-12-08 2017-04-18 Lg Electronics Inc. Device and method for receiving digital broadcast signal
US20140125780A1 (en) * 2010-12-08 2014-05-08 L G Electronics Inc. Device and method for receiving digital broadcast signal
US20130250051A1 (en) * 2010-12-13 2013-09-26 Electronics And Telecommunications Research Institute Signaling method for a stereoscopic video service and apparatus using the method
US20120206568A1 (en) * 2011-02-10 2012-08-16 Google Inc. Computing device having multiple image capture devices and image modes
US11122253B2 (en) 2011-05-24 2021-09-14 Tivo Corporation Dynamic distribution of multi-dimensional multimedia content
US10368052B2 (en) * 2011-05-24 2019-07-30 Comcast Cable Communications, Llc Dynamic distribution of three-dimensional content
US8988453B2 (en) * 2011-06-17 2015-03-24 Lg Display Co., Ltd. Stereoscopic image display device and driving method thereof
US20120320036A1 (en) * 2011-06-17 2012-12-20 Lg Display Co., Ltd. Stereoscopic Image Display Device and Driving Method Thereof
US9407872B2 (en) 2011-06-24 2016-08-02 At&T Intellectual Property I, Lp Apparatus and method for managing telepresence sessions
US9270973B2 (en) 2011-06-24 2016-02-23 At&T Intellectual Property I, Lp Apparatus and method for providing media content
US9736457B2 (en) 2011-06-24 2017-08-15 At&T Intellectual Property I, L.P. Apparatus and method for providing media content
US8947497B2 (en) 2011-06-24 2015-02-03 At&T Intellectual Property I, Lp Apparatus and method for managing telepresence sessions
US9681098B2 (en) 2011-06-24 2017-06-13 At&T Intellectual Property I, L.P. Apparatus and method for managing telepresence sessions
US10033964B2 (en) 2011-06-24 2018-07-24 At&T Intellectual Property I, L.P. Apparatus and method for presenting three dimensional objects with telepresence
US9602766B2 (en) 2011-06-24 2017-03-21 At&T Intellectual Property I, L.P. Apparatus and method for presenting three dimensional objects with telepresence
US9445046B2 (en) 2011-06-24 2016-09-13 At&T Intellectual Property I, L.P. Apparatus and method for presenting media content with telepresence
US10200669B2 (en) 2011-06-24 2019-02-05 At&T Intellectual Property I, L.P. Apparatus and method for providing media content
US10200651B2 (en) 2011-06-24 2019-02-05 At&T Intellectual Property I, L.P. Apparatus and method for presenting media content with telepresence
US9030522B2 (en) 2011-06-24 2015-05-12 At&T Intellectual Property I, Lp Apparatus and method for providing media content
US9160968B2 (en) 2011-06-24 2015-10-13 At&T Intellectual Property I, Lp Apparatus and method for managing telepresence sessions
US10484646B2 (en) 2011-06-24 2019-11-19 At&T Intellectual Property I, L.P. Apparatus and method for presenting three dimensional objects with telepresence
US9167205B2 (en) 2011-07-15 2015-10-20 At&T Intellectual Property I, Lp Apparatus and method for providing media services with telepresence
US9414017B2 (en) 2011-07-15 2016-08-09 At&T Intellectual Property I, Lp Apparatus and method for providing media services with telepresence
US9807344B2 (en) 2011-07-15 2017-10-31 At&T Intellectual Property I, L.P. Apparatus and method for providing media services with telepresence
US8587635B2 (en) 2011-07-15 2013-11-19 At&T Intellectual Property I, L.P. Apparatus and method for providing media services with telepresence
US20150089564A1 (en) * 2012-04-23 2015-03-26 Lg Electronics Inc. Signal processing device and method for 3d service
US9780910B2 (en) * 2013-03-14 2017-10-03 Harris Corporation Systems and methods for multiple stream encoded digital video
US20140269937A1 (en) * 2013-03-14 2014-09-18 Harris Corporation Systems and methods for multiple stream encoded digital video
US20140340490A1 (en) * 2013-05-15 2014-11-20 Paul Duffy Portable simulated 3d projection apparatus

Also Published As

Publication number Publication date
CN102474635B (en) 2017-02-08
CN102474635A (en) 2012-05-23
EP2454886A4 (en) 2013-01-30
KR20120041764A (en) 2012-05-02
EP2454886A1 (en) 2012-05-23
WO2011008917A1 (en) 2011-01-20
US9036700B2 (en) 2015-05-19
KR101342294B1 (en) 2013-12-16
EP2454886B1 (en) 2018-11-14

Similar Documents

Publication Publication Date Title
US9036700B2 (en) Simulcast of stereoviews for 3D TV
US10051275B2 (en) Methods and apparatus for encoding video content
US9554198B2 (en) Digital broadcast receiving method providing two-dimensional image and 3D image integration service, and digital broadcast receiving device using the same
JP6382329B2 (en) Broadcast signal transmission and reception method and apparatus for panorama service
KR20120026026A (en) Broadcast receiver and 3d video data processing method thereof
US8982186B2 (en) Method for providing and recognizing transmission mode in digital broadcasting
US20120154534A1 (en) Cable broadcast receiver and 3d video data processing method thereof
KR101844227B1 (en) Broadcast receiver and method for processing 3d video data
US8483271B2 (en) Method and system for 3D video pre-processing and post-processing
US9930382B2 (en) Method and apparatus for transmitting/receiving broadcast signal for 3-dimensional (3D) broadcast service
KR20150066301A (en) Method and apparatus for transmitting and receiving of high-definition contents
KR20150037200A (en) An apparatus of transmitting/receiving a video stream and a method of transmitting/receiving the video stream thereof
Kwon et al. Program associated 3D non-real-time service on terrestrial DTV
Lee et al. High-definition 3DTV trial service over terrestrial broadcasting network
KR101277267B1 (en) Coding method and apparatus for 3D broadcasting
KR20150086801A (en) Methods and apparatuses for transmitting and receiving additional video data for improvement of image quality
KR20120087869A (en) Coding method and apparatus for 3D broadcasting
KR20120139643A (en) Coding method and apparatus for 3d broadcasting

Legal Events

Date Code Title Description
AS Assignment

Owner name: GENERAL INSTRUMENT CORPORATION, PENNSYLVANIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LUTHRA, AJAY K.;NARASIMHAN, MANDAYAM A.;SIGNING DATES FROM 20100806 TO 20100811;REEL/FRAME:024897/0857

AS Assignment

Owner name: MOTOROLA MOBILITY LLC, ILLINOIS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GENERAL INSTRUMENT HOLDINGS, INC.;REEL/FRAME:030866/0113

Effective date: 20130528

Owner name: GENERAL INSTRUMENT HOLDINGS, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GENERAL INSTRUMENT CORPORATION;REEL/FRAME:030764/0575

Effective date: 20130415

AS Assignment

Owner name: GOOGLE TECHNOLOGY HOLDINGS LLC, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MOTOROLA MOBILITY LLC;REEL/FRAME:034301/0001

Effective date: 20141028

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8