WO2015198553A1 - データ出力装置、データ出力方法及びデータ生成方法 - Google Patents
データ出力装置、データ出力方法及びデータ生成方法 Download PDFInfo
- Publication number
- WO2015198553A1 WO2015198553A1 PCT/JP2015/002973 JP2015002973W WO2015198553A1 WO 2015198553 A1 WO2015198553 A1 WO 2015198553A1 JP 2015002973 W JP2015002973 W JP 2015002973W WO 2015198553 A1 WO2015198553 A1 WO 2015198553A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- metadata
- hdr
- luminance
- data
- video signal
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 146
- 230000005540 biological transmission Effects 0.000 claims abstract description 41
- 238000006243 chemical reaction Methods 0.000 claims description 262
- 230000003068 static effect Effects 0.000 claims description 100
- 238000012545 processing Methods 0.000 claims description 41
- 230000000153 supplemental effect Effects 0.000 claims description 5
- 238000010586 diagram Methods 0.000 description 63
- 230000009977 dual effect Effects 0.000 description 42
- 230000008569 process Effects 0.000 description 28
- 230000006870 function Effects 0.000 description 15
- 230000008859 change Effects 0.000 description 13
- 238000009826 distribution Methods 0.000 description 10
- 230000000694 effects Effects 0.000 description 9
- 238000005516 engineering process Methods 0.000 description 9
- 238000013139 quantization Methods 0.000 description 9
- 238000004590 computer program Methods 0.000 description 4
- 210000003127 knee Anatomy 0.000 description 4
- 238000003672 processing method Methods 0.000 description 4
- 238000003860 storage Methods 0.000 description 4
- 238000007792 addition Methods 0.000 description 3
- 239000000470 constituent Substances 0.000 description 3
- 238000004519 manufacturing process Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 101000997616 Mycobacterium tuberculosis (strain ATCC 25618 / H37Rv) 4-hydroxy-3-methylbut-2-enyl diphosphate reductase 1 Proteins 0.000 description 1
- 101001055850 Mycobacterium tuberculosis (strain ATCC 25618 / H37Rv) 4-hydroxy-3-methylbut-2-enyl diphosphate reductase 2 Proteins 0.000 description 1
- 230000002411 adverse Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- FFBHFFJDDLITSX-UHFFFAOYSA-N benzyl N-[2-hydroxy-4-(3-oxomorpholin-4-yl)phenyl]carbamate Chemical compound OC1=C(NC(=O)OCC2=CC=CC=C2)C=CC(=C1)N1CCOCC1=O FFBHFFJDDLITSX-UHFFFAOYSA-N 0.000 description 1
- LHJQIRIGXXHNLA-UHFFFAOYSA-N calcium peroxide Chemical compound [Ca+2].[O-][O-] LHJQIRIGXXHNLA-UHFFFAOYSA-N 0.000 description 1
- 230000001771 impaired effect Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/57—Control of contrast or brightness
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/10—Intensity circuits
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B20/00—Signal processing not specific to the method of recording or reproducing; Circuits therefor
- G11B20/10—Digital recording or reproducing
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
- G11B27/30—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
- G11B27/3027—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording used signal is digitally coded
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/436—Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/436—Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
- H04N21/43622—Interfacing an external recording device
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/436—Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
- H04N21/4363—Adapting the video stream to a specific local network, e.g. a Bluetooth® network
- H04N21/43632—Adapting the video stream to a specific local network, e.g. a Bluetooth® network involving a wired protocol, e.g. IEEE 1394
- H04N21/43635—HDMI
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/14—Picture signal circuitry for video frequency region
- H04N5/20—Circuitry for controlling amplitude response
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/775—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television receiver
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/08—Systems for the simultaneous or sequential transmission of more than one television signal, e.g. additional information signals, the signals occupying wholly or partially the same frequency band, e.g. by time division
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/64—Circuits for processing colour signals
- H04N9/68—Circuits for processing colour signals for controlling the amplitude of colour signals, e.g. automatic chroma control circuits
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2320/00—Control of display operating conditions
- G09G2320/06—Adjustment of display parameters
- G09G2320/0673—Adjustment of display parameters for control of gamma adjustment, e.g. selecting another gamma curve
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/04—Changes in size, position or resolution of an image
- G09G2340/0407—Resolution change, inclusive of the use of different resolutions for different screen areas
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2340/00—Aspects of display data processing
- G09G2340/04—Changes in size, position or resolution of an image
- G09G2340/0407—Resolution change, inclusive of the use of different resolutions for different screen areas
- G09G2340/0428—Gradation resolution change
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2370/00—Aspects of data communication
- G09G2370/04—Exchange of auxiliary data, i.e. other than image data, between monitor and graphics controller
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2370/00—Aspects of data communication
- G09G2370/12—Use of DVI or HDMI protocol in interfaces along the display data pipeline
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/003—Details of a display terminal, the details relating to the control arrangement of the display terminal and to the interfaces thereto
- G09G5/005—Adapting incoming signals to the display format of the display terminal
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/003—Details of a display terminal, the details relating to the control arrangement of the display terminal and to the interfaces thereto
- G09G5/006—Details of the interface to the display terminal
Definitions
- This disclosure relates to a data output device, a data output method, and a data generation method.
- Patent Document 1 An image signal processing apparatus for improving the displayable luminance level has been disclosed (see, for example, Patent Document 1).
- a data output device acquires a first metadata regarding a luminance range of the first video signal, and a decoding unit that generates a first video signal of the first luminance range by decoding the video stream
- An interpreter that acquires characteristic data indicating a luminance range of the first video signal by interpreting the first metadata, and the characteristic data is converted into control information according to a predetermined transmission protocol.
- a control information generating unit for converting, and an output unit for outputting the control information by the transmission protocol.
- a data generation method is a data generation method performed by a data generation device, the first generation step of generating metadata related to a luminance range of a video signal, the video signal, and the metadata
- FIG. 1 is a diagram for explaining the evolution of video technology.
- FIG. 2 is a diagram for explaining the relationship between the master, the distribution method, and the display device when HDR is introduced.
- FIG. 3 is an explanatory diagram of a method for determining the code value of the luminance signal stored in the content and the process of restoring the luminance value from the code value during reproduction.
- FIG. 4 is a diagram illustrating an example of HDR metadata.
- FIG. 5 is a diagram illustrating a storage example of static HDR metadata.
- FIG. 6 is a diagram illustrating a storage example of dynamic HDR metadata.
- FIG. 7 is a flowchart of a method for transmitting static HDR metadata.
- FIG. 8 is a flowchart of an HDR metadata processing method.
- FIG. 1 is a diagram for explaining the evolution of video technology.
- FIG. 2 is a diagram for explaining the relationship between the master, the distribution method, and the display device when HDR is introduced.
- FIG. 3 is an explanatory diagram
- FIG. 9 is a block diagram showing the configuration of the data output apparatus.
- FIG. 10 is a diagram illustrating an example of a data structure of an SEI message storing HDR metadata.
- FIG. 11 is a diagram illustrating an example of a data structure of an SEI message storing HDR metadata.
- FIG. 12 is a diagram illustrating an example of a user guidance display method.
- FIG. 13 is a diagram illustrating an example of a user guidance display method.
- FIG. 14 is a diagram illustrating an example of a user guidance display method.
- FIG. 15 is a diagram illustrating an example of a user guidance display method.
- FIG. 16 is a diagram for explaining the reproduction operation of the dual disc.
- FIG. 17 is a flowchart showing the dual disk playback operation.
- FIG. 10 is a diagram illustrating an example of a data structure of an SEI message storing HDR metadata.
- FIG. 11 is a diagram illustrating an example of a data structure of an SEI message storing HDR metadata
- FIG. 18 is a diagram showing the types of BDs.
- FIG. 19 is a diagram illustrating the types of BDs in more detail.
- FIG. 20 is a first diagram showing the data capacity recorded on the BD.
- FIG. 21 is a second diagram showing the data capacity recorded on the BD.
- FIG. 22 is a diagram illustrating an example of a combination of a video stream and a graphic stream recorded on each of a BD and a dual stream disc.
- FIG. 23 is a diagram showing another example of a combination of a video stream and a graphic stream recorded on each of a BD disc and a dual stream disc.
- FIG. 24 is a diagram showing still another example of a combination of a video stream and a graphic stream recorded on each of a BD and a dual stream disc.
- FIG. 20 is a first diagram showing the data capacity recorded on the BD.
- FIG. 21 is a second diagram showing the data capacity recorded on the BD.
- FIG. 22 is a diagram illustrating an
- FIG. 25A is a diagram illustrating an example of a display process in which HDR display is performed by converting an HDR signal in HDRTV.
- FIG. 25B is a diagram illustrating an example of a display process for performing HDR display using an HDR-compatible playback device and SDRTV.
- FIG. 25C is a diagram illustrating an example of display processing for performing HDR display on the HDR-compatible playback device and SDRTV that are connected to each other via the standard interface.
- FIG. 26 is a diagram for explaining the conversion process from HDR to pseudo-HDR.
- FIG. 27A is a diagram illustrating an example of EOTF (Electro-Optical Transfer Function) corresponding to each of HDR and SDR.
- FIG. 27B is a diagram illustrating an example of inverse EOTF corresponding to each of HDR and SDR.
- FIG. 28 is a block diagram illustrating configurations of the conversion device and the display device according to the embodiment.
- FIG. 29 is a flowchart illustrating a conversion method and a display method performed by the conversion device and the display device according to the embodiment.
- FIG. 30A is a diagram for describing the first luminance conversion.
- FIG. 30B is a diagram for describing another example of the first luminance conversion.
- FIG. 31 is a diagram for describing the second luminance conversion.
- FIG. 32 is a diagram for describing the third luminance conversion.
- FIG. 33 is a flowchart showing detailed display setting processing.
- a data output device acquires a first metadata regarding a luminance range of the first video signal, and a decoding unit that generates a first video signal of the first luminance range by decoding the video stream
- An interpreter that acquires characteristic data indicating a luminance range of the first video signal by interpreting the first metadata, and the characteristic data is converted into control information according to a predetermined transmission protocol.
- a control information generating unit for converting, and an output unit for outputting the control information by the transmission protocol.
- the data output device can generate control information based on the characteristic data included in the metadata.
- the interpretation unit further obtains conversion auxiliary data for converting a luminance range of the first video signal by interpreting the first metadata
- the data output device further includes the conversion A conversion unit configured to generate a second video signal having a second luminance range narrower than the first luminance range by converting a luminance range of the first video signal based on auxiliary data
- the output unit further includes: At least one of the first video signal and the second video signal may be output using the transmission protocol.
- the data output device can change the luminance range of the first video signal by using the conversion auxiliary data included in the metadata.
- the decoding unit further obtains second metadata related to a luminance range of the first video signal from the video stream, and the first metadata or the second metadata is a continuous reproduction unit of the video signal. It is used in common for a plurality of included images and includes static metadata including the characteristic data, and the control information generation unit converts the characteristic data included in the static metadata according to the transmission protocol.
- the output information may be converted into the control information, and the output unit may output the control information based on the static metadata when outputting the first video signal.
- the first metadata or the second metadata is further commonly used for a unit smaller than a continuous reproduction unit of a video signal, includes dynamic metadata including the characteristic data, and generates the control information.
- the unit converts the characteristic data included in the static metadata and the characteristic data included in the dynamic metadata into the control information according to the transmission protocol, and the output unit includes the first video.
- the control information based on the static metadata and the dynamic metadata may be output.
- the decoding unit further acquires second metadata relating to a luminance range of the first video signal from the video stream, and the interpreting unit receives at least one of the first metadata and the second metadata.
- the characteristic data and the conversion auxiliary data may be acquired by analysis.
- the first metadata and the second metadata are static metadata commonly used for a plurality of images included in a continuous reproduction unit of the first video signal, and the interpretation unit includes the interpretation unit
- the characteristic data and the conversion auxiliary data may be acquired by analyzing the first metadata
- the first metadata and the second metadata are static metadata commonly used for a plurality of images included in a continuous reproduction unit of the first video signal
- the interpretation unit includes the interpretation unit
- the characteristic data and the conversion auxiliary data are acquired by analyzing the first metadata
- the second metadata is acquired.
- the metadata to be used may be switched from the first metadata to the second metadata.
- the output unit displays the first video signal and the control information on the display device. May be output.
- the output unit may determine whether or not the display device supports video output in the luminance range of the first video signal in the initialization process of the transmission protocol.
- the output unit may output the control information in synchronization with a second video signal in which the first metadata is valid.
- the first metadata includes type information indicating a type of the first metadata
- the interpretation unit can interpret the first metadata by using the type information. If the data output device can interpret the first metadata, the characteristic data and the conversion auxiliary data may be acquired by interpreting the first metadata.
- the first metadata is stored in an SEI (Supplemental Enhancement Information) message of HEVC (High Efficiency Video Coding), and the SEI message identifies whether the SEI message includes metadata.
- the output unit receives an SEI message including the identification information indicating that the SEI message includes metadata, and the data output destination display device responds to the input of the control information. If so, the SEI message may be output as it is according to the transmission protocol.
- the output unit outputs the second video signal and the control information to the display device when the display device does not support video output in the luminance range of the first video signal, and
- the display device may display that the second video signal converted from the luminance range to the second luminance range is displayed.
- the output unit when the display device does not support the video output of the luminance range of the first video signal, the output unit outputs the second video signal converted from the first luminance range to the second luminance range.
- a message for the user to select whether to display or not may be displayed on the display device.
- the output unit includes a plurality of conversion methods for converting the first luminance range into the second luminance range when the display device does not support the video output of the luminance range of the first video signal. A message for the user to select one of them may be displayed on the display device.
- a data output method is a data output method performed by a data output device, wherein a decoding step of generating a first video signal in a first luminance range by decoding a video stream; An acquisition step of acquiring first metadata relating to a luminance range of the first video signal; an interpretation step of acquiring characteristic data indicating the luminance range of the first video signal by interpreting the first metadata; A control information generating step of converting the characteristic data into control information according to a predetermined transmission protocol; and an output step of outputting the control information by the transmission protocol.
- the data output method can generate control information based on the characteristic data included in the metadata.
- a data generation method is a data generation method performed by a data generation device, the first generation step of generating metadata related to a luminance range of a video signal, the video signal, and the metadata
- the data generation method can generate a video stream including metadata including characteristic data.
- the control information can be generated based on the characteristic data included in the metadata.
- the metadata may include a basic part whose interpretation is essential and an extension part whose interpretation is arbitrary.
- the metadata may include a payload, information indicating whether the data of the payload is a basic part or an extension part, and information indicating a data size of the payload.
- the metadata is stored in an SEI (Supplemental Enhancement Information) message of HEVC (High Efficiency Video Coding), and the SEI message may include identification information indicating whether the SEI message includes metadata. Good.
- SEI Supplemental Enhancement Information
- HEVC High Efficiency Video Coding
- each of the embodiments described below shows a specific example of the present disclosure.
- constituent elements that are not described in the independent claims indicating the highest concept are described as optional constituent elements.
- An HDR (High Dynamic Range) signal which is an image signal having a luminance range higher than that of a conventional image signal, is a package media such as a Blu-ray (registered trademark, hereinafter the same) disc storing the HDR signal, broadcast, or OTT (Over Thee). It is distributed via a distribution medium such as (Top).
- OTT means a Web site provided on the Internet, content or service such as video or audio, or a provider that provides them.
- the distributed HDR signal is decoded by a Blu-ray device or the like.
- the decoded HDR signal is sent to an HDR compatible display device (TV, projector, tablet, smartphone, etc.), and an HDR video is reproduced by the HDR compatible display device.
- the HDR technology is still at an early stage, and it is assumed that a new HDR method will be developed after the introduction of the first introduced HDR technology.
- the new HDR method can be adopted by storing the HDR signal (and metadata) of the newly created HDR method in the HDR distribution medium.
- the original technology can be used without changing a decoding device (for example, a Blu-ray device) designed for the original distribution medium.
- a decoding device for example, a Blu-ray device
- FIG. 1 is a diagram for explaining the evolution of video technology.
- SD Standard Definition
- HD high definition 1920 x 1080 pixels
- the dynamic range is the maximum brightness to express bright light such as specular reflection light that cannot be expressed by the current TV signal with more realistic brightness while maintaining the dark gradation in the conventional video.
- HDR High Dynamic Range
- SDR Standard Dynamic Range
- the maximum luminance value was 100 nits, whereas in HDR the maximum is 1000 nits or more. It is assumed that the luminance value is enlarged. Standardization of HDR is underway in SMPTE (Society of Motion Picture & Television Engineers) and ITU-R (International Telecommunications Union Radiocommunications Sector).
- HDR high definition video recorder
- package media Blu-ray Disc, etc.
- Internet distribution etc., like HD and UHD.
- the luminance of the video is composed of luminance values in the HDR luminance range, and a luminance signal obtained by quantizing the luminance value of the video is referred to as an HDR signal.
- the luminance of the video is composed of luminance values in the luminance range of SDR, and a luminance signal obtained by quantizing the luminance value of the video is called an SDR signal.
- FIG. 2 is a diagram showing a relationship between a flow for producing a master for home entertainment of SDR and HDR, a distribution medium, and a display device.
- the HDR concept has been proposed and its effectiveness at the HDR concept level has been confirmed.
- the first implementation method of HDR is proposed.
- a large amount of HDR content was created using this method, and the first implementation method was not verified. For this reason, when the production of HDR content becomes full-scale in the future, there is a possibility that the current HDR production method, particularly metadata, will change.
- FIG. 3 is an explanatory diagram of a method for determining the code value of the luminance signal stored in the content and the process of restoring the luminance value from the code value during reproduction.
- the luminance signal indicating the luminance in this example is an HDR signal corresponding to HDR.
- the image after grading is quantized by the inverse EOTF of HDR, and the code value corresponding to the luminance value of the image is determined. Image coding or the like is performed based on this code value, and a video stream is generated. At the time of reproduction, the decoding result of the stream is converted into a linear signal by inverse quantization based on HDR EOTF, and the luminance value for each pixel is restored.
- quantization using the inverse EOTF of HDR is referred to as “inverse HDR EOTF conversion”.
- Inverse quantization using HDR EOTF is referred to as “HDR EOTF conversion”.
- quantization using inverse SDR EOTF is referred to as “inverse SDR EOTF conversion”.
- Inverse quantization using SDR EOTF is referred to as “SDR EOTF conversion”.
- the HDR image can be displayed on the video display unit by converting the luminance value and metadata into a luminance value that can be displayed on the video display unit by the video conversion processing unit. For example, when the peak brightness of the original HDR video is 2000 nits and the peak brightness of the video display unit is 800 nits, conversion can be performed to lower the brightness.
- the HDR master method is realized by a combination of EOTF, metadata, and HDR signal. Therefore, more efficient EOTF and metadata may be developed, and it may be time to adopt the HDR method using such EOTF and metadata.
- This disclosure aims to promote the spread of HDR by reducing the risk that a customer who bought an HDR-compatible device re-buys a new device even when the HDR transmission format is changed in this way.
- FIG. 4 is a diagram illustrating an example of HDR metadata.
- the HDR metadata includes conversion auxiliary information used for changing the luminance range of the video signal (DR conversion) and HDR control information.
- Each piece of information is either static HDR metadata provided in units of titles or dynamic HDR metadata provided in units of frames, for example.
- static HDR metadata is classified as either essential metadata (basic data) or selection metadata (extended data), and dynamic HDR metadata is classified as selection metadata. Details of each information will be described later.
- HDR metadata As parameters indicating characteristics at the time of mastering in HDR content, there are static HDR metadata that is fixed for each title or each playlist and dynamic HDR metadata that is variable for each scene.
- the title and the playlist are information indicating video signals that are continuously reproduced.
- video signals that are continuously played back are referred to as continuous playback units.
- static HDR metadata includes at least one of EOTF function (curve) type, 18% Gray value, Diffuse White value, Knee point, and Clip point.
- the EOTF is information that associates a plurality of luminance values with a plurality of code values, and is information for changing the luminance range of the video signal. Since the other information is attribute information regarding the luminance of the video signal, the static HDR metadata is information regarding the luminance range of the video signal and can be said to be information for specifying the luminance range of the video signal.
- the 18% Gray value and the Diffuse White value indicate the luminance value (nit) in a video with a predetermined reference brightness, in other words, the reference brightness in the video. More specifically, the 18% Gray value indicates a luminance value (nit) after mastering of an object having a brightness of 18 nits before mastering.
- the Diffuse White value indicates a luminance value (nit) corresponding to white.
- Knee point and Clip point are parameters of the EOTF function and indicate the point at which the characteristics of EOTF change. Specifically, Knee point differs from the one-to-one increment in the luminance value (output luminance) mapped to the EOTF as the luminance of the video signal with respect to the original luminance value (input luminance) at the time of shooting. Indicates the change point to be a value. For example, Knee point is information for specifying a point that deviates from a linear change in FIG. 30A described later.
- Clip point indicates the point at which clipping starts in the EOTF function.
- clip refers to converting an input luminance value greater than a certain value into the same output luminance value. For example, Clip point indicates a point where the output luminance value does not change in FIG. 30B described later.
- EOTF functions are, for example, HDR EOTF and SDR EOTF shown in FIG. 27A.
- the content data generation method is a content data generation method for generating content data, in which a video signal and a plurality of images (continuous playback units are included in the continuous playback unit of the video signal)
- the information regarding the luminance range of the video signal is information for converting the luminance range of the video signal.
- the static HDR metadata includes information for specifying an EOTF that associates a plurality of luminance values with a plurality of code values.
- the luminance value in the video signal is encoded as a code value.
- the static HDR metadata further includes information indicating a luminance value in a video signal having a predetermined reference brightness, or information indicating a point at which a characteristic in EOTF changes.
- the static HDR metadata includes information (Diffuse White value) indicating a luminance value corresponding to white in the video signal.
- dynamic HDR metadata (second metadata) that is information that is commonly used for a unit smaller than the continuous reproduction unit and that is information on the luminance range of the video signal is further generated.
- the information regarding the luminance range of the video signal is information for converting the luminance range of the video signal.
- the dynamic HDR metadata is a parameter indicating a mastering characteristic that is different for each scene.
- the mastering characteristic indicates the relationship between the original luminance (before mastering) and the luminance after mastering.
- the parameter indicating the mastering characteristic is the same information as the static HDR metadata described above, in other words, is at least one piece of information included in the static HDR metadata.
- FIG. 5 is a diagram showing a storage example of static HDR metadata.
- static HDR metadata is stored in a playlist in a package medium such as a Blu-ray disc.
- Static HDR metadata is stored as one piece of metadata for each stream referenced from the playlist.
- static HDR metadata is fixed for each playlist. That is, the static HDR metadata is stored in association with each playlist.
- static HDR metadata may be stored in a manifest file that is referred to prior to stream acquisition. That is, the content data generation method according to the present embodiment may generate a video signal as a video stream, and store static HDR metadata in a manifest file that is referenced prior to acquisition of the video stream.
- static HDR metadata may be stored in a descriptor indicating the attribute of a stream. That is, the content data generation method according to the present embodiment may generate content data as a video stream, and store static HDR metadata as an identifier indicating the attribute of the video stream independently of the video stream.
- static HDR metadata can be stored as a descriptor (descriptor) in MPEG2-TS.
- the static HDR metadata when the static HDR metadata is fixed for each title, the static HDR metadata may be stored as management information indicating title attributes.
- FIG. 6 is a diagram illustrating an example of storing dynamic HDR metadata in a video stream.
- MPEG-4 AVC or HEVC High Efficiency Video Coding
- information related to stream reproduction control is stored using a data structure called SEI (Supplemental Enhancement Information).
- SEI Supplemental Enhancement Information
- dynamic HDR metadata is stored in SEI.
- the head of the scene is a head access unit (AU) of a random access unit such as GOP (Group Of Pictures). Therefore, the dynamic HDR metadata may be stored in the head access unit in the decoding order in the random access unit.
- the head access unit of the random access unit is an IDR picture or a non-IDR I picture to which SPS (Sequence Parameter Set) is added. Therefore, the receiving-side apparatus can acquire dynamic HDR metadata by detecting a NAL (Network Abstraction Layer) unit that constitutes a first access unit in a random access unit.
- NAL Network Abstraction Layer
- SEI that stores dynamic HDR metadata.
- the type of the EOTF function may be stored as stream attribute information in the SPS. That is, in the content data generation method according to the present embodiment, content data may be generated as a video stream encoded by HEVC, and information for specifying EOTF may be stored in an SPS included in the video stream.
- FIG. 7 is a diagram illustrating a transmission method of static HDR metadata.
- a playback device such as a BD player (Blu-ray device) or a recorder
- the display device transmits the static HDR metadata to the display device through a transmission protocol such as HDMI (registered trademark, the same applies hereinafter).
- HDMI registered trademark, the same applies hereinafter
- the playback device acquires the static HDR metadata from the content management information at the start of playback of the title or playlist.
- Static HDR metadata is stored and transmitted as HDMI control information.
- the playback device acquires static HDR metadata corresponding to the title or playlist prior to the start of transmission of the video signal constituting the title or playlist, and controls the acquired static HDR metadata according to HDMI. It transmits as information (S402). More generally, the playback device may transmit static HDR metadata as initialization information when performing an HDMI initialization process between the playback device and the display device.
- the playback device transmits a video stream corresponding to the static HDR metadata (S403). Note that the transmitted static HDR metadata is valid for this video stream.
- the video stream transmission method is a video stream transmission method for transmitting a video stream (video stream), and is common to a video signal and a plurality of images included in a continuous reproduction unit.
- content data including static HDR metadata (first metadata) relating to a luminance range of a video signal, a video stream corresponding to the video signal, and a static HDR meta A transmission step of transmitting data.
- the video stream and static HDR metadata are transmitted according to the HDMI communication protocol.
- dynamic HDR metadata is transmitted as part of the video stream.
- the playback apparatus may transmit the dynamic HDR metadata as an HDMI control signal at a timing when the dynamic HDR metadata becomes valid.
- the playback device transmits the static HDR metadata and the dynamic HDR metadata so that they can be distinguished from each other by providing an identifier or the like.
- only the data structure of the container for storing the dynamic HDR metadata may be defined so that the contents of the SEI can be copied as it is as the payload data of the container.
- the reproduction device can be changed even if the syntax of the static HDR metadata is changed. It can be handled without changing the implementation. That is, the container data structure for storing the static HDR metadata is defined. In the transmission step, the static HDR metadata included in the content data is copied to the payload of the container, and the container is transmitted. May be.
- FIG. 8 is a flowchart illustrating an example of a method for processing HDR metadata when an HDR signal is displayed on the display device.
- the display device acquires static HDR metadata from the HDMI control information (S411), and determines a display method of the HDR signal based on the acquired static HDR metadata (S412).
- the display device determines an HDR signal display method based on a value predetermined in the application standard or a default setting of the display device. . That is, in the video display method according to the present embodiment, when static HDR metadata cannot be acquired, the video display method corresponding to the video signal is determined based on a predetermined value or setting.
- the display device updates the HDR signal display method based on the dynamic HDR metadata (S414). That is, in the video display method according to the present embodiment, when static HDR metadata is acquired, the display method is determined based on the acquired static HDR metadata, the video is displayed, and dynamic HDR metadata is acquired. In such a case, the display method determined based on the static HDR metadata is updated to the display method determined based on the dynamic HDR metadata, and the video is displayed. Alternatively, the display method may be determined based on both static HDR metadata and dynamic HDR metadata.
- the display device may operate based only on static HDR metadata. Even when the display device supports the acquisition of dynamic HDR metadata, the display device displays the HDR signal in synchronization with the display time (PTS: Presentation Time Stamp) of the access unit storing the metadata. The method may not be updated. In this case, the display device may update the display method from the access unit displayed after the earliest time at which the display method can be updated after obtaining the metadata.
- PTS Presentation Time Stamp
- each of the static HDR metadata and the dynamic HDR metadata may include a plurality of versions, and may include a basic unit that is used in common by a plurality of versions and an extension unit that is different for each version. By doing so, backward compatibility in the display device can be ensured based on the HDR metadata of the basic part.
- the video display method according to the present embodiment is a video display method for displaying video based on the video stream, and the video stream corresponding to the video signal and the static HDR metadata (first metadata). And a display step of determining and displaying a video display method corresponding to the video signal based on the static HDR metadata.
- the luminance value in the video signal is encoded as a code value
- the static HDR metadata includes information for specifying an EOTF that associates a plurality of luminance values with a plurality of code values, and is displayed.
- a video is generated by converting a code value indicated by the video signal into a luminance value using EOTF specified by the static HDR metadata.
- FIG. 9 is a block diagram showing a configuration of a data output device 400 that outputs an HDR signal such as a BD player.
- the HDR metadata input to the data output device 400 indicates characteristic data indicating the mastering characteristic of the HDR signal and a tone mapping method when converting the HDR signal into the SDR signal or converting the dynamic range of the HDR signal. Conversion auxiliary data.
- These two types of metadata are stored as static HDR metadata or dynamic HDR metadata, as described with reference to FIGS.
- the static HDR metadata is stored in at least one of the content management information and the video stream.
- the data output device 400 includes a video decoding unit 401, an external meta acquisition unit 402, an HDR meta interpretation unit 403, an HDR control information generation unit 404, a DR conversion unit 405, and an HDMI output unit 406.
- the video decoding unit 401 generates a video signal (first video signal) by decoding a video stream that is an encoded video stream, and outputs the obtained video signal to the DR conversion unit 405.
- the video decoding unit 401 acquires HDR metadata (second metadata) (static HDR metadata or dynamic HDR metadata) in the video stream.
- the video decoding unit 401 outputs the HDR metadata stored in an MPEG-4 AVC or HEVC SEI message to the HDR meta interpretation unit 403.
- the external metadata acquisition unit 402 acquires static HDR metadata (first metadata) stored in content management information such as a playlist, and outputs the acquired static HDR metadata to the HDR meta interpretation unit 403. .
- static HDR metadata first metadata
- dynamic HDR metadata that can be changed in a predetermined unit that can be randomly accessed, such as a play item, may be stored in the content management information.
- the external meta acquisition unit 402 acquires dynamic HDR metadata from content management information, and outputs the acquired dynamic HDR metadata to the HDR meta interpretation unit 403.
- the HDR meta interpretation unit 403 determines the type of HDR metadata output from the video decoding unit 401 or the external meta acquisition unit 402, outputs the characteristic data to the HDR control information generation unit 404, and converts the conversion auxiliary data into the DR conversion unit. Output to 405.
- the HDR meta interpretation unit 403 obtains the characteristic data and the conversion auxiliary data by analyzing the first metadata when the first metadata and the second metadata are obtained together. To do.
- the HDR meta interpretation unit 403 uses the static HDR metadata as valid metadata when the external metadata acquisition unit 402 acquires the static HDR metadata, and the video decoding unit 401 further uses the static HDR metadata.
- valid metadata may be overwritten with the static HDR metadata. That is, the first metadata acquired by the external metadata acquisition unit 402 and the second metadata acquired by the video decoding unit 401 are commonly used for a plurality of images included in the continuous playback unit of the first video signal.
- the HDR meta interpretation unit 403 analyzes the first metadata when only the first metadata is acquired from the first metadata and the second metadata. When the data and the conversion auxiliary data are acquired and the second metadata is acquired, the metadata to be used is switched from the first metadata to the second metadata.
- the HDR control information generation unit 404 generates HDR control information in HDMI based on the characteristic data, and outputs the generated HDR control information to the HDMI output unit 406.
- the output timing of the HDR control information in the HDMI output unit 406 is determined so that the HDR control information can be output in synchronization with the video signal for which the metadata is valid. That is, the HDMI output unit 406 outputs the HDR control information in synchronization with a video signal (video signal) for which metadata becomes valid.
- the DR conversion unit 405 converts the decoded video signal into an SDR signal or converts the dynamic range based on the conversion auxiliary data.
- the data output device 400 can determine whether conversion processing is necessary by confirming whether the connected display device supports the input of an HDR signal in an HDMI initialization process or the like. Good.
- the first video signal obtained by the video decoding unit 401 is input to the HDMI output unit 406 without passing through the DR conversion unit 405.
- the HDMI output unit 406 displays the first video signal and the HDR control information when the display device connected to the data output device 400 supports video output in the luminance range of the HDR signal (first video signal). Output to the display device. Also, the HDMI output unit 406 displays the second video obtained by converting HDR into SDR when the display device connected to the data output device 400 does not support video output in the luminance range of the HDR signal (first video signal). The signal and the HDR control information are output to the display device. Also, the HDMI output unit 406 determines whether or not the display device supports video output in the luminance range of the HDR signal (first video signal) in the initialization process of the transmission protocol (for example, HDMI).
- the transmission protocol for example, HDMI
- the HDMI output unit 406 outputs the video signal output from the DR conversion unit 405 or the video decoding unit 401 and the HDR control information according to the HDMI protocol.
- the same configuration can be used when the data output device 400 receives and outputs broadcast or OTT content. Further, when the data output device 400 and the display device are included in a single device, the HDMI output unit 406 is not necessary.
- the data output device 400 includes the external metadata acquisition unit 402 that acquires metadata from management information and the like, and the video decoding unit 401 has a function of acquiring metadata from a video stream.
- the output device 400 may have only one of them.
- the data output device 400 outputs data (video signal and HDR control information) according to HDMI has been described. However, if the data output device 400 outputs data according to an arbitrary transmission protocol, Good.
- the data output device 400 relates to the decoding unit (video decoding unit 401) that generates the first video signal in the first luminance range (HDR) by decoding the video stream, and the luminance range of the first video signal.
- the acquisition unit at least one of the video decoding unit 401 and the external metadata acquisition unit 402 and interpreting the first metadata acquire characteristic data indicating the luminance range of the first video signal.
- An interpretation unit HDR meta-interpretation unit 403
- a control information generation unit HDR control information generation unit 404 that converts characteristic data into HDR control information according to a predetermined transmission protocol (for example, HDMI), and HDR control information
- an output unit (HDMI output unit 406) that outputs a predetermined transmission protocol.
- the data output device 400 can generate control information based on the characteristic data included in the metadata.
- the interpretation unit (HDR meta-interpretation unit 403) further obtains conversion auxiliary data for converting the luminance range of the first video signal by interpreting the first metadata
- the data output device 400 Furthermore, a conversion unit (DR conversion unit 405) is provided that generates a second video signal having a luminance range narrower than the luminance range of the first video signal by converting the luminance range of the first video signal based on the auxiliary conversion data.
- the output unit (HDMI output unit 406) further outputs at least one of the first video signal and the second video signal using a predetermined transmission protocol.
- the data output device 400 can change the luminance range of the first video signal by using the conversion auxiliary data included in the metadata.
- the decoding unit (video decoding unit 401) further acquires second metadata (HDR metadata) related to the luminance range of the first video signal from the video stream, and the interpretation unit (HDR meta-interpretation unit 403)
- the characteristic data and the conversion auxiliary data are acquired by analyzing at least one of the first metadata and the second metadata.
- static HDR metadata includes essential metadata and selection metadata
- dynamic HDR metadata includes only selection metadata. That is, static HDR metadata is always used, and dynamic HDR metadata is selectively used.
- the first metadata acquired by the external meta acquisition unit 402 or the second metadata acquired by the video decoding unit 401 is used in common for a plurality of images included in the continuous reproduction unit of the video signal.
- static HDR metadata static metadata
- the HDR control information generation unit 404 converts the characteristic data included in the static HDR metadata into HDR control information according to a predetermined transmission protocol.
- the HDMI output unit 406 outputs HDR control information based on static HDR metadata.
- the first metadata acquired by the external metadata acquisition unit 402 or the second metadata acquired by the video decoding unit 401 is further commonly used for a unit smaller than the continuous reproduction unit of the video signal, and has characteristics.
- Dynamic HDR metadata including data (dynamic metadata) is included.
- the HDR control information generation unit 404 converts the characteristic data included in the static HDR metadata and the specific data included in the dynamic HDR metadata into HDR control information according to a predetermined transmission protocol.
- the HDMI output unit 406 outputs HDR control information based on static HDR metadata and dynamic HDR metadata.
- a data generation method is a data generation method performed by a data generation device, and includes a first generation step of generating metadata related to a luminance range of a video signal, and a video stream including the video signal and metadata.
- a second generation step of generating The metadata includes characteristic data indicating the luminance range of the video signal and conversion auxiliary data for converting the luminance range of the video signal.
- FIG. 10 is a diagram illustrating an example of a data structure of an SEI message storing HDR metadata.
- a dedicated SEI message for HDR metadata may be defined. That is, the metadata may be stored in a message dedicated to metadata.
- HDR metadata is stored in a general-purpose SEI message for storing user data, and information (HDR extended identification information described later) indicating that HDR metadata is stored in the message in the payload portion of the message. It may be provided.
- the HDR metadata includes static HDR metadata and dynamic HDR metadata. Further, flag information indicating whether static HDR metadata is stored and flag information indicating whether dynamic HDR metadata is stored may be provided.
- there are three types of storage a method of storing only static HDR metadata, a method of storing only dynamic HDR metadata, and a method of storing both static HDR metadata and dynamic HDR metadata. The method can be used.
- basic data (basic part) that must be interpreted and extended data (extended part) that is optional (interpretation is optional)
- type information indicating the type of metadata (basic data or extension data) and a size are included in the header information, and a container format in which the metadata is stored in the payload is defined. That is, the metadata includes a payload, information indicating whether the payload data is basic data or extension data, and information indicating the payload data size.
- the metadata includes type information indicating the type of metadata. For example, basic data is stored in a container whose type value is 0. Further, one or more values are assigned as type values to the extension data, and the type of extension data is indicated by the value.
- the data output device and the display device refer to the type value and acquire container data that can be interpreted by the data output device and the display device.
- the data output device uses the type information to determine whether the data output device (or display device) can interpret the metadata, and the data output device (or display device) determines the metadata. If the data can be interpreted, the characteristic data and the conversion auxiliary data are acquired by interpreting the metadata.
- the maximum size of the HDR metadata may be set in advance, and the metadata may be generated so that the sum of the sizes of the basic data and the extended data is not more than the maximum size. That is, the maximum value of the metadata data size is defined, and the data generation method according to the present disclosure generates the metadata so that the total data size of the basic data and the extended data is equal to or less than the maximum value.
- the HDR data can be stored in the memory by providing the data for the maximum size in the data output device and the display device.
- Such a data structure may be used for storing HDR metadata in content management information.
- FIG. 11 is a diagram illustrating an example of a data structure when HDR metadata is stored in an SEI message for storing user data.
- the data structure is the same as that of FIG. 11 except that the message includes the HDR extension identification information and the extension type ID.
- the HDR extended identification information indicates that the message includes HDR metadata.
- the extension type ID indicates the HDR metadata version and the like. That is, the metadata is stored in the HEVC SEI message, and the SEI message includes HDR extended identification information indicating whether the SEI message includes metadata.
- the data output device receives the SEI message for storing user data including the HDR extended identification information, and the display device connected to the data output device supports the input of the HDR signal and the HDR control information.
- the received SEI message is copied and output as it is according to the protocol of the output I / F to the display device such as HDMI. That is, the data output device acquires an SEI message including HDR extended identification information indicating that the SEI message includes metadata, and the data output destination display device supports input of HDR control information.
- the SEI message is output as it is according to a predetermined transmission protocol (for example, HDMI).
- the data output device can output the HDR metadata to the display device regardless of the content of the metadata.
- a new DR conversion process is developed in the future, new HDR metadata is defined, and a display device corresponding to the new HDR metadata is connected to a data output device not corresponding to the new HDR metadata. Even in this case, the new HDR metadata can be output from the data output device to the display device.
- DR conversion processing according to the new HDR metadata can be performed in the display device.
- FIG. 12 is a diagram showing a user guidance display method on a Blu-ray device that executes an HDR to SDR conversion process.
- an HDR-compatible Blu-ray device connected to a non-HDR compatible TV detects the start of conversion processing from HDR to SDR, for example, “The disk is an HDR-compatible disk. Your TV is a non-HDR compatible TV. Therefore, a guide message such as “SDR video in which the Blu-ray device performs the conversion process from HDR to SDR is reproduced instead of the HDR video” is displayed.
- the data output device converts the first luminance range to the second luminance range when the display device does not support the video output of the luminance range of the first video signal (HDR signal).
- the second video signal (SDR signal) and the HDR control information are output to the display device, and the display device displays that the second video signal converted from the first luminance range to the second luminance range is displayed.
- FIG. 13 is a diagram showing a method for displaying user guidance when executing a conversion process from HDR to SDR stored in the disc.
- a message (menu) to be displayed by the Blu-ray device is stored in the HDR disk or a non-volatile memory in the Blu-ray device.
- the Blu-ray device can display a message when executing the conversion process from HDR to SDR.
- the disc is an HDR-compatible disc. Because your TV is non-HDR compatible, the Blu-ray device plays SDR video that has been converted from HDR to SDR instead of HDR video. Is displayed.
- FIG. 14 is a diagram showing a display method of the user guidance menu when executing conversion processing from HDR to SDR stored in the disc.
- the Blu-ray devices use the Blu-ray menu, “The disc is an HDR-compatible disc. Because your TV is non-HDR compatible, the Blu-ray device is not HDR video, and the Blu-ray device is from HDR to SDR. Message such as “Is it possible to play the SDR video that has been subjected to the conversion process” is displayed? The Blu-ray device starts displaying the converted image when the user selects the “Play” button. Also, if the user selects “do not play”, the Blu-ray device stops playing and displays a message prompting the user to insert a non-HDR compatible Blu-ray disc.
- the data output device converts the first luminance range to the second luminance range when the display device does not support the video output of the luminance range of the first video signal (HDR signal).
- a message for the user to select whether or not to display the second video signal (SDR signal) is displayed on the display device.
- FIG. 15 is a diagram showing a display method of a user guidance menu capable of selecting a processing method when executing a conversion process from HDR to SDR stored in the disc.
- the Blu-ray device displays this information when metadata for conversion processing from HDR to SDR is stored in Blu-ray.
- the Blu-ray device displays a message prompting that a more beautiful conversion is possible when the user selects a specified conversion method.
- the Blu-ray device says, “The disc is an HDR-compatible disc. Because your TV is non-HDR compatible, the Blu-ray device is not an HDR video, and the Blu-ray device has converted HDR to SDR.
- Display the selection menu of the HDR to SDR conversion processing method such as “Play back in process 1”, “Play back in process 3”, “Do not play back”. be able to.
- processing 1 and processing 3 are different conversion processing from HDR to SDR.
- the data output device converts the first luminance range to the second luminance range when the display device does not support the video output of the luminance range of the first video signal (HDR signal).
- the display device displays a message for the user to select one of a plurality of conversion methods.
- a similar message can be displayed in broadcasting.
- a TV or playback device that does not support HDR signals uses a data broadcast application or the like to display a message indicating that the broadcast program is an HDR signal and may not be displayed correctly when viewed with To do.
- the TV or the playback device that supports the HDR signal may not display the message.
- the tag value indicating the message attribute indicates that the message is a warning message for the HDR signal. The TV or playback device corresponding to the HDR signal determines that display of the message is unnecessary with reference to the tag value.
- FIG. 16 is a diagram for explaining multiplexed data stored in a dual disk.
- the HDR signal and the SDR signal are stored as different multiplexed streams.
- data of a plurality of media such as video, audio, subtitles, and graphics is stored as one multiplexed stream by an MPEG-2 TS-based multiplexing method called M2TS.
- M2TS MPEG-2 TS-based multiplexing method
- These multiplexed streams are referred to from reproduction control metadata such as a playlist, and are reproduced by the player analyzing the metadata at the time of reproduction, or individual languages stored in the multiplexed stream. Select the data.
- playlists for HDR and SDR are individually stored, and each playlist refers to an HDR signal or an SDR signal. Further, identification information indicating that both the HDR signal and the SDR signal are stored may be separately indicated.
- HDR and SDR signals are multiplexed to satisfy a buffer model such as T-STD (System Target Decoder) defined in MPEG-2 TS.
- T-STD System Target Decoder
- Data such as audio, subtitles, or graphics needs to be stored for each multiplexed stream, and the amount of data increases compared to the case of multiplexing to one.
- the increase in the amount of data can reduce the amount of video data by using a video encoding method with a high compression rate.
- the compression ratio can be improved by 1.6 to 2 times.
- the dual disk is stored in a combination of 2K HDR and SDR, such as a combination of 4K SDR and 2K HDR, or 2K, or a combination of 2K and 4K. It is also possible to allow only combinations that fit within the capacity of the optical disk by prohibiting storing two.
- FIG. 17 is a flowchart showing the dual disk playback operation.
- the playback apparatus determines whether the playback target optical disk is a dual disk (S301). If it is determined that the disk is a dual disk (Yes in S301), it is determined whether the output TV is HDRTV or SDRTV (S302). If it is determined that it is HDRTV (Yes in S302), the process proceeds to step S303, and if it is determined that it is SDRTV (No in S302), the process proceeds to step S304. In step S303, an HDR video signal is acquired from the multiplexed stream including the HDR signal in the dual disc, decoded, and output to HDRTV.
- step S304 an SDR video signal is acquired from the multiplexed stream including the SDR signal in the dual disc, decoded, and output to SDRTV. If it is determined in step S301 that the playback target is not a dual disc (No in S301), whether or not playback is possible is determined by a predetermined method, and a playback method is determined based on the determination result (S305).
- FIG. 18 is a diagram showing the types of BDs.
- a BD on which a video signal having a resolution of the first resolution and a luminance range of the first luminance range is recorded as a 2K_SDR-compatible BD.
- a video signal whose resolution is the first resolution and whose luminance range is the first luminance range is stored as a stream on the BD. This stream is described as a 2K_SDR stream.
- the 2K_SDR compatible BD is a conventional BD.
- a BD on which a video signal having a resolution of the second resolution and a luminance range of the first luminance range is recorded as a 4K_SDR compatible BD.
- a video signal whose resolution is the second resolution and whose luminance range is the first luminance range is stored as a stream on the BD. This stream is described as a 4K_SDR stream.
- a BD on which a video signal having a resolution of the first resolution and a luminance range of the second luminance range is recorded as a 2K_HDR compatible BD.
- a video signal whose resolution is the first resolution and whose luminance range is the second luminance range is stored as a stream on the BD. This stream is described as a 2K_HDR stream.
- a BD on which a video signal having a resolution of the second resolution and a luminance range of the second luminance range is recorded as a 4K_HDR compatible BD.
- a video signal whose resolution is the second resolution and whose luminance range is the second luminance range is stored as a stream on the BD. This stream is described as a 4K_HDR stream.
- the first resolution is, for example, a so-called 2K resolution (1920 ⁇ 1080, 2048 ⁇ 1080), but may be any resolution including such a resolution.
- the first resolution may be simply described as 2K.
- the second resolution is a so-called 4K resolution (3840 ⁇ 2160, 4096 ⁇ 2160), but may be an arbitrary resolution including such a resolution.
- the second resolution is a resolution having more pixels than the first resolution.
- the first luminance range is, for example, the SDR (luminance range with a peak luminance of 100 nit) described so far.
- the second luminance range is, for example, the HDR described so far (luminance range in which the peak luminance exceeds 100 nits).
- the second luminance range includes the entire first luminance range, and the peak luminance of the second luminance range is larger than the peak luminance of the first luminance range.
- FIG. 19 is a diagram showing the types of BDs in more detail.
- a dual stream disc corresponding to a plurality of video representations with one BD can be considered.
- the dual stream disc is a BD on which a plurality of video signals for reproducing the same content and having a plurality of video signals different in at least one of resolution and luminance range are recorded.
- the dual stream disc shown in (c) of FIG. 19 is a BD in which a 4K_SDR stream and a 2K_SDR stream are recorded.
- the dual stream disc shown in (f) of FIG. 19 is a BD in which a 2K_HDR stream and a 2K_SDR stream are recorded.
- the dual stream disc shown in (g) of FIG. 19 is a BD on which a 4K_HDR stream and a 4K_SDR stream are recorded.
- the dual stream disc shown in (h) of FIG. 19 is a BD in which a 4K_HDR stream and a 2K_SDR stream are recorded.
- the dual stream disc shown in FIG. 19 (c) is not essential because the Blu-ray device can perform 4K to 2K resolution down-conversion (hereinafter also referred to as down-conversion).
- each BD as described above will be supplemented with reference to FIGS. 20 and 21 are diagrams showing the data capacity recorded on the BD.
- FIG. 20 shows an example in which a 2K resolution stream (2K_SDR stream and 2K_HDR stream) is compressed using MPEG-4 AVC.
- the bit rates of Movie length, lossless Audio, and Compressed Audio are as follows.
- the BD records audio streams (Lossless Audio and Compressed Audio) for the number of languages.
- the above dual stream disks can be realized in terms of capacity.
- FIG. 21 illustrates a case where streams with a resolution of 2K (2K_SDR stream and 2K_HDR stream) are compressed using HEVC.
- the bit rates of Movie length, lossless audio, and compressed audio are as follows.
- the above dual stream disks can be realized in terms of capacity.
- the disk capacity required for a dual stream disc on which both a 4K_HDR stream compressed using HEVC and a 2K_SDR stream compressed using HEVC are based on (a) + ( ⁇ ). 88.9 GB, 67.5 GB based on (b) + ( ⁇ ), 61.3 GB based on (b) + ( ⁇ ), and (c) + ( ⁇ ) Based on this, it is 57.3 GB. Therefore, such a dual stream disk can be realized by a disk having a capacity of 100 GB or a disk having a capacity of 66 GB.
- FIG. 22 is a diagram showing an example of a combination of a video stream and a graphic stream recorded on each disc for each BD including a dual stream disc.
- FIG. 22 shows that the graphic stream is recorded at a resolution of 2K regardless of the resolution of the corresponding video stream in consideration of the production of the content (BD).
- the graphic stream can be shared between the 2K_SDR stream and the 4K_SDR stream. However, the graphic stream is recorded in a luminance range that matches the luminance range of the corresponding video stream.
- the video stream is HDR
- an HDR graphics stream is recorded.
- the video stream is SDR
- an SDR graphics stream is recorded. Conversion of the graphic stream from SDR to HDR is performed at the time of content creation.
- FIG. 23 is a diagram showing another example of a combination of a video stream and a graphic stream recorded on each disc for each BD including a dual stream disc.
- the graphic stream is recorded with a resolution of 2K and a luminance range of SDR regardless of the resolution and luminance range of the corresponding video stream.
- the 2K_SDR stream, 4K_SDR stream, 2K_HDR stream, and 4K_HDR stream can all share a graphic stream.
- conversion of the resolution of the graphic stream from 2K to 4K and conversion of the luminance range of the graphic stream from SDR to HDR are both executed by the Blu-ray device.
- FIG. 24 is a diagram showing still another example of a combination of a video stream and a graphic stream recorded on each disc for each BD including a dual stream disc.
- FIG. 24 shows that the resolution and luminance range of the graphic stream are recorded in accordance with the resolution and luminance range of the corresponding video stream when producing the content so that the conversion of the graphic stream is not necessary in the Blu-ray device.
- a Blu-ray device that plays back a 4K compatible BD or HDR compatible BD needs to support four TVs: a 2K_SDR compatible TV, a 2K_HDR compatible TV, a 4K_SDR compatible TV, and a 4K_HDR compatible TV.
- the Blu-ray device needs to support three sets of HDMI / HDCP standards (HDMI 1.4 / HDCP 1.4, HDMI 2.0 / HDCP 2.1, HDMI 2.1 / HDCP 2.2).
- each BD (content) and connection It is necessary to select an appropriate process and HDMI / HDCP for each display device (TV). Furthermore, when a graphic is combined with a video, it is necessary to change processing depending on the type of BD and the type of display device (TV) connected thereto.
- the graphic stream is restricted and the types of combinations of the video stream and the graphic stream are reduced.
- the HDR video when HDR video is displayed in SDRTV, the HDR video is converted to SDR video of 100 nit or less by using the peak brightness of SDRTV to be displayed exceeds 100 nit (usually 200 nit or more). Instead, it realizes a “HDR ⁇ pseudo HDR conversion process” that can be converted to maintain a certain level of gradation in a region exceeding 100 nits, converted into pseudo HDR video close to the original HDR, and displayed on SDRTV.
- the conversion method of “HDR ⁇ pseudo HDR conversion processing” may be switched depending on the display characteristics (maximum luminance, input / output characteristics, and display mode) of SDRTV.
- the display characteristic information can be acquired by (1) automatic acquisition through HDMI or a network, (2) generation by allowing the user to input information such as manufacturer name and product number, and (3) manufacturer name and product number. It is conceivable to obtain information from the cloud using this information.
- the display characteristic information acquisition timing of the conversion device 100 includes (1) acquisition immediately before pseudo-HDR conversion, and (2) when connecting to the display device 200 (such as SDRTV) for the first time (when connection is established). ) Can be considered.
- the conversion method may be switched according to the luminance information (CAL, CPL) of the HDR video.
- the method of acquiring the luminance information of the HDR video of the conversion device 100 (1) acquiring as meta information attached to the HDR video, (2) acquiring by causing the user to input content title information, And (3) It is conceivable to acquire from the cloud or the like using input information that has been instructed by the user.
- the details of the conversion method are as follows: (1) Conversion so as not to exceed DPL; (2) Conversion so that CPL becomes DPL; (3) CAL and its surrounding luminance are not changed; 4) Conversion using natural logarithm, (5) Clip processing by DPL.
- display settings such as the SDRTV display mode and display parameters can be transmitted to the display device 200 and switched. For example, the user is prompted to perform display settings. A message may be displayed on the screen.
- FIG. 25A is a diagram illustrating an example of a display process in which HDR display is performed by converting an HDR signal in HDRTV.
- the maximum value of the HDR luminance range (peak luminance (HPL (HDR Peak Luminance): example 1500 nit)) is displayed as it is. May not be possible.
- the linear signal after inverse quantization using HDR EOTF is adjusted to the maximum value of the luminance range of the display device (peak luminance (DPL (Display Peak Iluminance): example 750 nit)).
- DPL Display Peak Iluminance
- FIG. 25B is a diagram illustrating an example of a display process for performing HDR display using an HDR-compatible playback device and SDRTV.
- FIG. 25C is a diagram illustrating an example of a display process for performing HDR display using an HDR-compatible playback device and SDRTV that are connected to each other via a standard interface.
- a signal input via an input interface passes through “SDR EOTF conversion”, “brightness conversion for each mode”, and “display device” in order, and an image that matches the maximum luminance range of the display device. Is displayed. For this reason, a signal (pseudo HDR signal) that can cancel the “SDR EOTF conversion” and “brightness conversion for each mode” that passes immediately after the input interface in SDRTV in the HDR-compatible Blu-ray device.
- an input interface such as HDMI
- a normal SDRTV has an input signal of 100 nits, but has an ability to express an image of 200 nits or more according to the viewing environment (dark room: cinema mode, bright room: dynamic mode, etc.). However, since the upper limit of the luminance of the input signal to SDRTV was determined to be 100 nits, it was not possible to directly use this capability.
- HDR video in SDRTV When displaying HDR video in SDRTV, using the fact that the peak brightness of SDRTV to be displayed exceeds 100 nits (usually 200 nits or more), the HDR video is not converted to SDR video of 100 nits or less, but luminance exceeding 100 nits. “HDR ⁇ pseudo HDR conversion processing” is performed so as to maintain the gradation of the range to some extent. For this reason, it can be displayed on SDRTV as a pseudo HDR video close to the original HDR.
- An HDR signal transmitted by broadcasting, package media such as Blu-ray, and Internet distribution such as OTT is converted into a pseudo HDR signal by performing HDR-pseudo HDR conversion processing. As a result, it is possible to display the HDR signal as a pseudo HDR video with the existing SDRTV.
- FIG. 27A is a diagram showing an example of EOTF (Electro-Optical Transfer Function) corresponding to each of HDR and SDR.
- EOTF Electro-Optical Transfer Function
- EOTF is generally called a gamma curve, indicates the correspondence between code values and luminance values, and converts code values into luminance values. That is, EOTF is relationship information indicating a correspondence relationship between a plurality of code values and luminance values.
- FIG. 27B is a diagram illustrating an example of reverse EOTF corresponding to each of HDR and SDR.
- Inverse EOTF indicates the correspondence between the luminance value and the code value.
- the luminance value is quantized and converted into a code value. That is, inverse EOTF is relationship information indicating a correspondence relationship between a luminance value and a plurality of code values. For example, when the luminance value of a video corresponding to HDR is expressed by a 10-bit gradation code value, the luminance values in the HDR luminance range up to 10,000 nits are quantized and 1024 from 0 to 1023 Mapped to the integer value of.
- HDR EOTF EOTF corresponding to HDR
- HDR inverse EOTF an EOTF corresponding to SDR
- SDR EOTF an EOTF corresponding to SDR
- inverse EOTF of SDR an inverse EOTF corresponding to SDR
- the value (peak luminance) is 10,000 nits. That is, the HDR luminance range includes the entire SDR luminance range, and the HDR peak luminance is larger than the SDR peak luminance.
- the HDR luminance range is a luminance range obtained by expanding the maximum value from 100 nit, which is the maximum value of the SDR luminance range, to 10,000 nit.
- HDR EOTF and HDR inverse EOTF are, for example, SMPTE 2084 standardized by the American Film and Television Engineers Association (SMPTE).
- the luminance range from 0 nit to 100 nit which is the peak luminance described in FIGS. 27A and 27B may be described as the first luminance range.
- the luminance range described in FIGS. 27A and 27B from 0 nit to 10,000 nit which is the peak luminance may be described as the second luminance range.
- FIG. 28 is a block diagram illustrating configurations of the conversion device and the display device according to the embodiment.
- FIG. 29 is a flowchart illustrating a conversion method and a display method performed by the conversion device and the display device according to the embodiment.
- the conversion apparatus 100 includes an HDR EOTF conversion unit 101, a luminance conversion unit 102, an inverse luminance conversion unit 103, and an inverse SDR EOTF conversion unit 104.
- the display device 200 includes a display setting unit 201, an SDR EOTF conversion unit 202, a luminance conversion unit 203, and a display unit 204.
- a conversion method performed by the conversion apparatus 100 will be described with reference to FIG.
- the conversion method includes steps S101 to S104 described below.
- the HDR EOTF conversion unit 101 of the conversion apparatus 100 acquires an HDR video that has been subjected to reverse HDR EOTF conversion.
- the HDR EOTF conversion unit 101 of the conversion device 100 performs HDR EOTF conversion on the HDR signal of the acquired HDR video (S101).
- the HDR EOTF converter 101 converts the acquired HDR signal into a linear signal indicating a luminance value.
- HDR EOTF is, for example, SMPTE 2084.
- the luminance conversion unit 102 of the conversion device 100 performs first luminance conversion that converts the linear signal converted by the HDR EOTF conversion unit 101 using display characteristic information and content luminance information (S102). .
- a luminance value corresponding to the HDR luminance range (hereinafter referred to as “HDR luminance value”) is changed to a luminance value corresponding to the display luminance range (hereinafter referred to as “display luminance value”). Convert. Details will be described later.
- the HDR EOTF converter 101 functions as an acquisition unit that acquires the HDR signal as the first luminance signal indicating the code value obtained by quantizing the luminance value of the video.
- the HDR EOTF conversion unit 101 and the luminance conversion unit 102 determine the code value indicated by the HDR signal acquired by the acquisition unit based on the luminance range of the display (display device 200). It functions as a conversion unit that converts to a display luminance value corresponding to the luminance range of the display that is a maximum value (DPL) that is smaller than the value (HPL) and larger than 100 nits.
- DPL maximum value
- the HDR EOTF conversion unit 101 uses the acquired HDR signal and the HDR EOTF in step S101 to determine the HDR code value as the first code value indicated by the acquired HDR signal.
- An HDR luminance value associated with the HDR code value in the HDR EOTF is determined.
- the HDR signal is obtained by quantizing the luminance value of the video (content) using the HDR inverse EOTF that associates the luminance value in the HDR luminance range with a plurality of HDR code values.
- the code value of HDR is shown.
- step S102 the luminance conversion unit 102 determines a display luminance value corresponding to the luminance range of the display, which is associated with the HDR luminance value in advance in the HDR luminance value determined in step S101.
- the first luminance conversion is performed to convert the HDR luminance value corresponding to the luminance range of the display into the display luminance value corresponding to the luminance range of the display.
- the conversion apparatus 100 includes content luminance information including at least one of a maximum luminance value (CPL: Content Peak luminance) and an average luminance value (CAL: Content Average luminance) of video (content) before step S102. Is acquired as information relating to the HDR signal.
- CPL first maximum luminance value
- CAL is an average luminance value that is an average of luminance values for a plurality of images constituting an HDR video, for example.
- the conversion device 100 acquires the display characteristic information of the display device 200 from the display device 200 before step S102.
- the display characteristic information refers to the maximum value (DPL) of luminance that can be displayed on the display device 200, the display mode of the display device 200 (see later), the input / output characteristics (EOTF corresponding to the display device), and the like. This is information indicating display characteristics.
- the conversion apparatus 100 may transmit recommended display setting information (referred to below-mentioned, hereinafter also referred to as “setting information”) to the display apparatus 200.
- recommended display setting information referred to below-mentioned, hereinafter also referred to as “setting information”.
- the reverse luminance conversion unit 103 of the conversion device 100 performs reverse luminance conversion according to the display mode of the display device 200. Accordingly, the inverse luminance conversion unit 103 performs the second luminance conversion for converting the luminance value corresponding to the luminance range of the display into the luminance value corresponding to the luminance range (0 to 100 [nit]) of the SDR (S103). . Details will be described later. That is, the inverse luminance conversion unit 103 uses the display luminance value obtained in step S102 as the third luminance value corresponding to the luminance range of SDR, which is associated with the display luminance value in advance and has a maximum value of 100 nits.
- SDR luminance value Luminance value corresponding to SDR
- the luminance value of SDR is determined, and the display luminance value corresponding to the luminance range of the display is changed to the luminance value of SDR corresponding to the luminance range of SDR.
- a second luminance conversion for conversion is performed.
- the inverse SDR EOTF converter 104 of the conversion device 100 performs pseudo SDR EOTF conversion to generate a pseudo HDR video (S104). That is, the inverse SDR EOTF conversion unit 104 performs the inverse dynamic EOTF (Electro-Optical) of SDR (Standard Dynamic Range), which is the third relational information that associates the luminance value in the HDR luminance range with a plurality of third code values.
- the determined SDR brightness value is quantized using Transfer Function), the third code value obtained by the quantization is determined, and the SDR brightness value corresponding to the SDR brightness range is indicated as the third code value.
- a pseudo HDR signal is generated by converting the SDR signal as the third luminance signal.
- the third code value is a code value corresponding to SDR, and is hereinafter referred to as “SDR code value”. That is, the SDR signal is obtained by quantizing the luminance value of the video using the SDR inverse EOTF that associates the luminance value in the luminance range of the SDR and a plurality of SDR code values. Expressed as a code value. Then, conversion device 100 outputs the pseudo HDR signal (SDR signal) generated in step S104 to display device 200.
- SDR code value a code value corresponding to SDR
- the conversion apparatus 100 generates the SDR luminance value corresponding to the pseudo HDR by performing the first luminance conversion and the second luminance conversion on the HDR luminance value obtained by dequantizing the HDR signal. Then, the SDR luminance value is quantized using the SDR EOTF to generate an SDR signal corresponding to the pseudo HDR.
- the SDR luminance value is a numerical value in the luminance range of 0 to 100 nits corresponding to the SDR.
- the HDR EOTF and the SDR are converted to the HDR luminance value. This is a numerical value different from the luminance value in the luminance range of 0 to 100 nit corresponding to the SDR obtained by performing luminance conversion using the EOTF.
- the display method includes steps S105 to S108 described below.
- the display setting unit 201 of the display device 200 sets the display setting of the display device 200 using the setting information acquired from the conversion device 100 (S105).
- the display device 200 is an SDRTV.
- the setting information is information indicating display settings recommended for the display device, and information indicating how to perform pseudo-HDR video EOTF and display at which setting a beautiful video can be displayed (that is, Information for switching the display setting of the display device 200 to the optimal display setting).
- the setting information includes, for example, a gamma curve characteristic at the time of output in the display device, a display mode such as a living mode (normal mode) and a dynamic mode, a numerical value of backlight (brightness), and the like.
- a message that prompts the user to change the display setting of the display device 200 by a manual operation may be displayed on the display device 200 (hereinafter also referred to as “SDR display”). Details will be described later.
- the display device 200 acquires an SDR signal (pseudo HDR signal) and setting information indicating display settings recommended for the display device 200 in displaying a video before step S105.
- the display device 200 may acquire the SDR signal (pseudo HDR signal) before step S106, or after step S105.
- the SDR EOTF converter 202 of the display device 200 performs SDR EOTF conversion on the acquired pseudo-HDR signal (S106). That is, the SDR EOTF converter 202 performs inverse quantization on the SDR signal (pseudo HDR signal) using the SDR EOTF. Accordingly, the SDR EOTF converter 202 converts the SDR code value indicated by the SDR signal into an SDR luminance value.
- the luminance conversion unit 203 of the display device 200 performs luminance conversion according to the display mode set in the display device 200. Thereby, the luminance conversion unit 203 converts the SDR luminance value corresponding to the SDR luminance range (0 to 100 [nit]) into the display luminance value corresponding to the display luminance range (0 to DPL [nit]). The third luminance conversion is performed (S107). Details will be described later.
- the display device 200 uses the setting information acquired in step S105 to obtain the third code value indicated by the acquired SDR signal (pseudo HDR signal) in step S106 and step S107. 0 to DPL [nit]).
- step S106 in the conversion from the SDR signal (pseudo HDR signal) to the display luminance value, in step S106, using the EOTF that associates the luminance value in the luminance range of the SDR with a plurality of third code values. Then, for the SDR code value indicated by the acquired SDR signal, the SDR luminance value related to the SDR code value by the SDR EOTF is determined.
- step S107 the display brightness value corresponding to the brightness range of the display, which is related in advance to the determined brightness value of the SDR, is determined, and the SDR value corresponding to the SDR brightness range is determined.
- a third luminance conversion is performed for converting the luminance value into a display luminance value corresponding to the luminance range of the display.
- the display unit 204 of the display device 200 displays the pseudo HDR video on the display device 200 based on the converted display luminance value (S108).
- FIG. 30A is a diagram for describing an example of the first luminance conversion.
- the luminance conversion unit 102 of the conversion device 100 performs first luminance conversion that converts the linear signal (HDR luminance value) obtained in step S101 using display characteristic information and content luminance information of the HDR video. .
- the HDR luminance value input luminance value
- the display luminance value output luminance value
- the DPL is determined using the maximum brightness and display mode of the SDR display, which is display characteristic information.
- the display mode is, for example, mode information such as a theater mode that is displayed dark on the SDR display and a dynamic mode that is displayed brightly.
- DPL second maximum luminance value
- DPL is the maximum luminance value that can be displayed in the display mode in which the SDR display is currently set. That is, in the first luminance conversion, DPL as the second maximum luminance value is determined using display characteristic information that is information indicating display characteristics of the SDR display.
- the first luminance conversion CAL and CPL in the content luminance information are used, luminance values below the CAL are the same before and after the conversion, and the luminance value is changed only for luminance values near the CPL.
- the first luminance conversion when the HDR luminance value is CAL or less, the HDR luminance value is not converted, and the HDR luminance value is determined as the display luminance value.
- the DPL as the second maximum luminance value is determined as the display luminance value.
- the peak luminance (CPL) of the HDR video in the luminance information is used, and when the HDR luminance value is CPL, DPL is determined as the display luminance value.
- the linear signal (HDR luminance value) obtained in step S101 may be converted so as to be clipped to a value not exceeding DPL.
- the processing in the conversion device 100 can be simplified, and the device can be reduced, the power consumption can be reduced, and the processing speed can be increased.
- FIG. 30B is a diagram for describing another example of the first luminance conversion.
- FIG. 31 is a diagram for describing the second luminance conversion.
- the inverse luminance conversion unit 103 of the conversion device 100 performs inverse luminance conversion corresponding to the display mode on the display luminance value in the display luminance range (0 to DPL [nit]) converted by the first luminance conversion in step S102. Apply.
- the reverse luminance conversion when the luminance conversion processing (step S107) according to the display mode by the SDR display is performed, the display luminance value of the display luminance range (0 to DPL [nit]) after the processing of step S102 is acquired. This is a process for making it possible. That is, the second luminance conversion is an inverse luminance conversion of the third luminance conversion.
- the second luminance conversion converts the display luminance value (input luminance value) in the luminance range of the display into the SDR luminance value (output luminance value) in the SDR luminance range.
- the conversion formula is switched depending on the display mode of the SDR display. For example, when the display mode of the SDR display is the normal mode, the luminance is converted to a directly proportional value that is directly proportional to the display luminance value.
- the second luminance conversion when the display mode of the SDR display is a dynamic mode in which a high luminance pixel is brighter and a low luminance pixel is darker than in the normal mode, the inverse function is used to obtain the low luminance pixel.
- the luminance value of the SDR is converted to a value higher than the direct proportional value directly proportional to the display luminance value, and the luminance value of the SDR of the high luminance pixel is converted to a value lower than the direct proportional value directly proportional to the display luminance value. That is, in the second luminance conversion, the display luminance value determined in step S102 is related to the display luminance value using luminance relationship information corresponding to the display characteristic information that is information indicating the display characteristic of the SDR display. The brightness value is determined as the brightness value of the SDR, and the brightness conversion process is switched according to the display characteristic information.
- the luminance-related information according to the display characteristic information includes, for example, a display luminance value (input luminance value) determined for each display parameter (display mode) of the SDR display as shown in FIG. This is information relating brightness values (output brightness values).
- FIG. 32 is a diagram for describing the third luminance conversion.
- the luminance conversion unit 203 of the display device 200 converts the SDR luminance value in the SDR luminance range (0 to 100 [nit]) into (0 to DPL [nit]) according to the display mode set in step S105. .
- This processing is performed so as to be an inverse function of inverse luminance conversion for each mode in S103.
- the conversion formula is switched depending on the display mode of the SDR display. For example, when the display mode of the SDR display is the normal mode (that is, when the set display parameter is a parameter corresponding to the normal mode), the display luminance value is converted into a direct proportional value that is directly proportional to the luminance value of the SDR. .
- the display mode of the SDR display is the dynamic mode in which the high luminance pixel is brighter and the low luminance pixel is darker than the normal mode, the display luminance value of the low luminance pixel is SDR.
- the luminance value of the display luminance value of the high luminance pixel is converted to a value that is higher than the directly proportional value that is directly proportional to the luminance value of the SDR, to a value lower than the directly proportional value that is directly proportional to the luminance value. That is, in the third luminance conversion, for the luminance value of the SDR determined in step S106, luminance related in advance to the luminance value of the SDR using luminance relationship information corresponding to the display parameter indicating the display setting of the SDR display. The value is determined as the display luminance value, and the luminance conversion processing is switched according to the display parameter.
- the luminance-related information according to the display parameter is, for example, as shown in FIG. 32, the SDR luminance value (input luminance value) determined for each display parameter (display mode) of the SDR display, and the display luminance. This is information that relates values (output luminance values).
- FIG. 33 is a flowchart showing detailed display setting processing.
- the display setting unit 201 of the SDR display performs the following steps S201 to S208 in step S105.
- the display setting unit 201 uses the setting information to determine whether the EOTF set for the SDR display (EOF for SDR display) matches the EOTF assumed when the pseudo HDR video (SDR signal) is generated. Determine (S201).
- the display setting unit 201 determines that the EOTF set in the SDR display is different from the EOTF indicated by the setting information (EOTF matching the pseudo HDR video) (Yes in S201), the display setting unit 201 sets the EOTF for the SDR display in the system. It is determined whether switching is possible on the side (S202).
- the display setting unit 201 determines that switching is possible, the setting information is used to switch the SDR display EOTF to an appropriate EOTF (S203).
- step S105 in the display setting setting (S105), the EOTF set in the SDR display is set to the recommended EOTF corresponding to the acquired setting information. Accordingly, in step S106 performed after step S105, the luminance value of SDR can be determined using the recommended EOTF.
- a message prompting the user to change the EOTF manually is displayed on the screen (S204). For example, a message “Please set the display gamma to 2.4” is displayed on the screen. That is, if the EOTF set for the SDR display cannot be switched in the display setting setting (S105), the display setting unit 201 switches the EOTF set for the SDR display (EOF for SDR display) to the recommended EOTF. A message for prompting the user to do so is displayed on the SDR display.
- the pseudo-HDR image (SDR signal) is displayed on the SDR display, but it is determined whether or not the display parameters of the SDR display match the setting information using the setting information before display (S205).
- the display setting unit 201 determines whether the display parameter of the SDR display can be switched (S206). .
- the display setting unit 201 determines that the display parameter of the SDR display can be switched (Yes in S206)
- the display setting unit 201 switches the display parameter of the SDR display according to the setting information (S207).
- step S105 in the display setting setting (S105), the display parameter set in the SDR display is set to the recommended display parameter corresponding to the acquired setting information.
- a message prompting the user to manually change the display parameters set in the SDR display is displayed on the screen (S208). For example, a message “Please set the display mode to dynamic mode and maximize the backlight” is displayed on the screen. That is, in the setting (S105), when the display parameter set in the SDR display cannot be switched, a message for prompting the user to switch the display parameter set in the SDR display to the recommended display parameter is displayed. To display.
- HDR video is, for example, Blu-ray Disc, DVD, Internet video distribution site, broadcast, video in HDD.
- the conversion device 100 may exist inside a disk player, a disk recorder, a set top box, a television, a personal computer, or a smartphone.
- the conversion device 100 may exist inside a server device in the Internet.
- the display device 200 (SDR display unit) is, for example, a television, a personal computer, or a smartphone.
- the display characteristic information acquired by the conversion device 100 may be acquired from the display device 200 via an HDMI cable or a LAN cable using HDMI or another communication protocol.
- the display characteristic information acquired by the conversion apparatus 100 may acquire display characteristic information included in the model information of the display apparatus 200 via the Internet.
- the user may perform a manual operation to set display characteristic information in the conversion device 100.
- the display characteristic information of the conversion device 100 may be acquired immediately before the pseudo HDR video generation (steps S101 to S104), or may be at the time of initial setting of the device or at the time of display connection.
- the display characteristic information may be acquired immediately before the conversion to the display luminance value, or may be performed at the timing when the conversion device 100 is first connected to the display device 200 with the HDMI cable.
- HDR video CPL or CAL there may be one HDR video CPL or CAL per content, or there may be one for each scene. That is, in the conversion method, the luminance information corresponding to each of the plurality of scenes of the video, and for each scene, the first maximum luminance value that is the maximum value among the luminance values for the plurality of images constituting the scene. And luminance information (CPL, CAL) including at least one of average luminance values that are averages of luminance values for a plurality of images constituting the scene, and in the first luminance conversion, for each of the plurality of scenes, The display brightness value may be determined according to the brightness information corresponding to the scene.
- CPL and CAL may be included in the same medium (Blu-ray Disc, DVD, etc.) as the HDR video, or acquired from a different location from the HDR video, such as acquired by the conversion device 100 from the Internet. May be. That is, luminance information including at least one of CPL and CAL may be acquired as video meta information or may be acquired via a network.
- a fixed value may be used without using CPL, CAL, and display peak luminance (DPL). Further, the fixed value may be changed from the outside.
- CPL, CAL, and DPL may be switched in several types. For example, DPL may be set to only three types of 200 nit, 400 nit, and 800 nit, and the value closest to the display characteristic information is used. You may make it do.
- the HDR EOTF may not be SMPTE 2084, but other types of HDR EOTF may be used.
- the maximum luminance (HPL) of the HDR video may not be 10,000 nits, for example, 4,000 nits or 1,000 nits.
- the bit width of the code value may be 16, 14, 12, 10, 8 bits, for example.
- inverse SDR EOTF conversion is determined from the display characteristic information, a fixed conversion function (which can be changed from the outside) may be used. Inverse SDR EOTF conversion is performed in, for example, Rec. ITU-R BT. A function defined in 1886 may be used. Also, the types of inverse SDR EOTF conversion may be limited to several types, and the one closest to the input / output characteristics of the display device 200 may be selected and used.
- the display mode may be a fixed mode and may not be included in the display characteristic information.
- the conversion device 100 may not transmit the setting information, and the display device 200 may have a fixed display setting or may not change the display setting.
- the display setting unit 201 is not necessary.
- the setting information may be flag information indicating whether or not the image is a pseudo HDR video.
- the setting information may be changed to a setting that displays the brightest image. That is, in the display setting setting (S105), when the acquired setting information indicates that the signal indicates a pseudo-HDR image converted using DPL, the brightness setting of the display device 200 is set to display the brightest. You may switch to.
- L indicates a luminance value normalized to 0 to 1
- S1, S2, a, b, and M are values set based on CAL, CPL, and DPL.
- V is a luminance value after conversion normalized to 0 to 1.
- CAL is set to 300 nit
- CPL is set to 2,000 nit
- DPL is set to 750 nit
- CAL + is not converted to 50 nit
- conversion is performed for 350 nit or more
- each value is, for example, It becomes such a value.
- the conversion formula By converting HDR video using information such as content peak brightness and content average brightness of the HDR video, the conversion formula can be changed according to the content, and conversion can be performed so as to keep the HDR gradation as much as possible. It becomes. In addition, adverse effects such as being too dark and too bright can be suppressed. Specifically, the gradation is kept as much as possible by mapping the content peak luminance of the HDR video to the display peak luminance. In addition, the overall brightness is prevented from changing by not changing the pixel values below the average luminance.
- the conversion formula can be changed according to the display environment of the SDR display, and there is a sense of HDR according to the performance of the SDR display.
- Video can be displayed with the same gradation and brightness as the original HDR video.
- the display peak brightness is determined according to the maximum brightness of the SDR display and the display mode, and the HDR video is converted so as not to exceed the peak brightness value.
- the display is performed with almost no gradation reduction, and the brightness that cannot be displayed is reduced to the displayable brightness.
- the display can be displayed in a form close to the original HDR video without reducing the gradation of displayable brightness.
- the overall brightness is maintained by converting to a pseudo HDR video with a peak luminance of 1,000 nits, and the luminance value changes depending on the display mode of the display. For this reason, the luminance conversion formula is changed according to the display mode of the display. If the pseudo-HDR image allows a luminance larger than the peak luminance of the display, the high luminance may be replaced with the peak luminance on the display side, and in that case, it is darker than the original HDR video. Become.
- the performance relating to the gradation of the display is not used at the maximum.
- the pseudo HDR video can be better displayed by switching the display setting using the setting information. For example, when the brightness is set to dark, high brightness display cannot be performed, so that the HDR feeling is impaired. In that case, by changing the display setting or displaying a message prompting the user to change the display setting, the display performance can be maximized and a high gradation video can be displayed.
- each component may be configured by dedicated hardware such as a circuit, or may be realized by executing a software program suitable for each component.
- Each component may be realized by a program execution unit such as a CPU or a processor reading and executing a software program recorded on a recording medium such as a hard disk or a semiconductor memory.
- the present disclosure can be applied to a content data generation device, a video stream transmission device such as a Blu-ray device, or a video display device such as a television.
- conversion device 101 EOTF conversion unit 102 luminance conversion unit 103 reverse luminance conversion unit 104 reverse SDR EOTF conversion unit 200 display device 201 display setting unit 202 SDR EOTF conversion unit 203 luminance conversion unit 204 display unit 400 data output device 401 video decoding Unit 402 external meta acquisition unit 403 HDR meta interpretation unit 404 HDR control information generation unit 405 DR conversion unit 406 HDMI output unit
Landscapes
- Engineering & Computer Science (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Computer Hardware Design (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
- Television Signal Processing For Recording (AREA)
- Controls And Circuits For Display Device (AREA)
- Transforming Electric Information Into Light Information (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
- Processing Of Color Television Signals (AREA)
Abstract
Description
[1.背景]
従来の画像信号より輝度範囲が高い画像信号であるHDR(High Dynamic Range)信号は、HDR信号を格納したBlu-ray(登録商標、以下同様)ディスク等のパッケージメディア、放送、又はOTT(Over The Top)等の配信媒体経由で配信される。ここで、OTTとは、インターネット上で提供されるWebサイト、動画或いは音声などのコンテンツ或いはサービス、又はそれらを提供する事業者を意味する。配信されたHDR信号は、Blu-ray機器等により復号される。また、復号されたHDR信号は、HDR対応表示装置(TV、プロジェクタ、タブレット、又はスマートフォン等)に送られ、HDR対応表示装置によりHDR映像が再生される。
図2は、SDRとHDRのホームエンターテイメント用マスターを制作するフローと、配信媒体及び表示装置の関係を示す図である。
図3は、コンテンツに格納される輝度信号のコード値の決定方法、および、再生時にコード値から輝度値を復元するプロセスの説明図である。
図4は、HDRメタデータの例を示す図である。HDRメタデータは、映像信号の輝度範囲の変更(DR変換)に用いられる変換補助情報と、HDR制御情報とを含む。各情報は、例えばタイトル単位で設けられる静的HDRメタデータと、例えばフレーム単位で設けられる動的HDRメタデータとのいずれかである。また、静的HDRメタデータは、必須メタデータ(基本データ)と、選択メタデータ(拡張データ)とのいずれかに分類され、動的HDRメタデータは、選択メタデータに分類される。なお、各情報の詳細は後述する。
HDRコンテンツにおけるマスタリング時の特性を示すパラメータとしては、タイトル毎又はプレイリスト毎に固定である静的HDRメタデータと、シーン毎に可変である動的HDRメタデータとが存在する。ここで、タイトル及びプレイリストは、連続して再生される映像信号を示す情報である。以降、連続して再生される映像信号を連続再生単位と呼ぶ。
図7は、静的HDRメタデータの伝送方法を示す図であり、BDプレーヤ(Blu-ray機器)又はレコーダなどの再生装置においてHDMI(登録商標、以下同様)などの伝送プロトコルを通じて、表示装置にHDR信号を伝送する際の動作例を示すフローチャートである。
図8は、表示装置においてHDR信号を表示する際のHDRメタデータの処理方法の例を示すフローチャートである。まず、表示装置は、HDMIの制御情報から静的HDRメタデータを取得し(S411)、取得した静的HDRメタデータに基づき、HDR信号の表示方法を決定する(S412)。
図9は、BDプレーヤなどのHDR信号を出力するデータ出力装置400の構成を示すブロック図である。データ出力装置400に入力されるHDRメタデータは、HDR信号のマスタリング特性を示す特性データと、HDR信号をSDR信号に変換する、又は、HDR信号のダイナミックレンジを変換する際のトーンマッピング方法を示す変換補助データとを含む。これら2種類のメタデータは、図5及び図6において説明したように、静的HDRメタデータ、又は、動的HDRメタデータとして格納される。さらに、静的HDRメタデータは、コンテンツの管理情報及び、映像ストリーム内の少なくとも一方に格納される。
図10は、HDRメタデータを格納するSEIメッセージのデータ構造例を示す図である。図10に示すように、HDRメタデータ専用のSEIメッセージが定義されてもよい。つまり、メタデータは、メタデータ専用のメッセージに格納されてもよい。
図12は、HDRからSDRへの変換処理を実行するBlu-ray機器でのユーザーガイダンス表示方法を示す図である。
図13は、ディスク内に格納されたHDRからSDRへの変換処理実行時のユーザーガイダンスの表示方法を示す図である。
図14は、ディスク内に格納されたHDRからSDRへの変換処理実行時のユーザーガイダンスメニューの表示方法を示す図である。
図15は、ディスク内に格納されたHDRからSDRへの変換処理実行時の処理方法を選択可能なユーザーガイダンスメニューの表示方法を示す図である。
なお、放送においても同様のメッセージを表示することができる。例えば、HDR信号に対応していないTV又は再生装置は、データ放送のアプリケーションなどを用いて、放送番組がHDR信号であり、で視聴した場合には正しく表示できないことがある旨を示すメッセージを表示する。また、HDR信号に対応したTV又は再生装置は、当該メッセージを表示しなくてもよい。また、メッセージの属性を示すタグ値などにより、当該メッセージがHDR信号に対する警告メッセージであることが示される。HDR信号に対応したTV又は再生装置は、タグ値を参照してメッセージの表示が不要であることを判定する。
以上では、HDR信号のみが格納されたHDRディスクの再生動作について説明した。
図17は、デュアルディスクの再生動作を示すフローチャートである。
上述のように、表示装置が高解像度化及び高輝度範囲化されることで、表示装置の仕様に合わせた複数種別のBlu-ray Disc(以下、BDと記載する)を提供する。図18は、BDの種類を示す図である。図18に示されるように、以下では、解像度が第1解像度であり、輝度範囲が第1輝度範囲である映像信号が記録されたBDは、2K_SDR対応BDと記載する。解像度が第1解像度であり、輝度範囲が第1輝度範囲である映像信号は、BDにストリームとして格納される。このストリームは、2K_SDRストリームと記載する。2K_SDR対応BDは、従来のBDはである。
ここで、以上説明したような各BDについて、図20及び図21を用いて補足する。図20及び図21は、BDに記録されるデータ容量を示す図である。
Lossless Audio: 0 -2 language (4.5mbps)
Compressed Audio: 3 -5 language (1.5mbps)
この場合、必要なディスク容量の最大値(A)、中間値(B)、及び最小値(C)は、以下のようになる。
(B)(16+4.5*1+1.5*3)mbps*(150*60)s/8 = 28.1 GB
(C)(14+4.5*0+1.5*3)mbps*(150*60)s/8 = 20.8 GB
また、解像度が4Kのストリーム(4K_SDRストリーム及び4K_HDRストリーム)が、HEVCを用いて圧縮されている場合を例示している。Movie length、lossless Audio、Compressed Audioのビットレートは、下記のようになる。
Lossless Audio: 0 -2 language (4.5mbps)
Compressed Audio: 3 -6 language (1.5mbps)
この場合、必要なディスク容量の最大値(a)、中間値(b)、及び最小値(c)は、以下のようになる。
(b)(37+4.5*0+1.5*4)mbps*(150*60)s/8 = 48.4 GB
(c)(35+4.5*0+1.5*3)mbps*(150*60)s/8 = 44.4 GB
ここで、MPEG-4 AVCを用いて圧縮された2K_HDRストリームと、MPEG-4 AVCを用いて圧縮された2K_SDRストリームとの両方が記録されたデュアルストリームディスクに必要なディスク容量は、上記(A)+(A)、(B)+(B)、及び、(C)+(C)により求められる。具体的には、最大値77.6GB、中間値56.2GB、及び、最小値41.6GBとなる。
さらに、図21を用いて別の例について説明する。図21は、解像度が2Kのストリーム(2K_SDRストリーム及び2K_HDRストリーム)が、HEVCを用いて圧縮されている場合を例示している。Movie length、lossless Audio、Compressed Audioのビットレートは、下記のようになる。
Lossless Audio: 0 - 2 language (4.5mbps)
Compressed Audio: 3 - 5 language (1.5mbps)
この場合、必要なディスク容量の最大値(A)、中間値(B)、及び最小値(C)は、以下のようになる。
(β)(8+4.5*1+1.5*3)mbps*(150*60)s/8 = 19.1 GB
(γ)(7+4.5*0+1.5*3)mbps*(150*60)s/8 = 12.9 GB
ここで、HEVCを用いて圧縮された2K_HDRストリームと、HEVCを用いて圧縮された2K_SDRストリームとの両方が記録されたデュアルストリームディスクに必要なディスク容量は、上記(α)+(α)、(β)+(β)、及び、(γ)+(γ)により求められる。具体的には、最大値50.6GB、typ値38.2GB、及び、最小値25.8GBとなる。
BDには、より詳細にはビデオストリームと、グラフィックストリーム(実施の形態1のグラフィックスのストリーム)とが記録される。ここで、図22は、デュアルストリームディスクを含む各BDに対し、各ディスクに記録された、ビデオストリームとグラフィックストリームとの組み合わせの一例を示す図である。
図23は、デュアルストリームディスクを含む各BDに対し、各ディスクに記録された、ビデオストリームとグラフィックストリームとの組み合わせの別の例を示す図である。
図24は、デュアルストリームディスクを含む各BDに対し、各ディスクに記録された、ビデオストリームとグラフィックストリームとの組み合わせのさらに別の例を示す図である。
4K対応BDまたはHDR対応BDを再生するBlu-ray機器は、2K_SDR対応TV、2K_HDR対応TV、4K_SDR対応TV、及び、4K_HDR対応TVの4つのTVに対応する必要がある。具体的には、Blu-ray機器は、3組のHDMI/HDCP規格(HDMI1.4/HDCP1.4、HDMI2.0/HDCP2.1、HDMI2.1/HDCP2.2)をサポートする必要がある。
次に、疑似HDRの必要性について図25A~図25Cを用いて説明する。
通常のSDRTVは入力信号が100nitであるが、視聴環境(暗い室:シネマモード、明るい部屋:ダイナミックモード等)に合わせて200nit以上の映像表現が可能な能力を持つ。しかし、SDRTVへの入力信号の輝度上限が100nitに決められていたため、その能力を直接つかうことはできなかった。
放送、Blu-ray等のパッケージメディア、OTT等のインターネット配信により送られてきたHDR信号を、HDR-疑似HDR変換処理を行うことで、疑似HDR信号に変換する。これにより、HDR信号を疑似HDR映像として既存のSDRTVで表示することが可能となる。
ここで、EOTFについて、図27Aおよび図27Bを用いて説明する。
図28は、実施の形態の変換装置および表示装置の構成を示すブロック図である。図29は、実施の形態の変換装置および表示装置により行われる変換方法および表示方法を示すフローチャートである。
変換装置100が行う変換方法について、図29を用いて説明する。なお、変換方法は、以下で説明するステップS101~ステップS104を含む。
次に、ステップS102の第1輝度変換(HPL→DPL)の詳細について、図30Aを用いて説明する。図30Aは、第1輝度変換の一例について説明するための図である。
次に、ステップS103の第2輝度変換(DPL→100〔nit〕)の詳細について、図31を用いて説明する。図31は、第2輝度変換について説明するための図である。
次に、ステップS107の第3輝度変換(100→DPL〔nit〕)の詳細について、図32を用いて説明する。図32は、第3輝度変換について説明するための図である。
次に、ステップS105の表示設定の詳細について、図33を用いて説明する。図33は、表示設定の詳細な処理を示すフローチャートである。
以上のように、本出願において開示する技術の例示として、実施の形態を説明した。しかしながら、本開示における技術は、これに限定されず、適宜、変更、置き換え、付加、省略などを行った実施の形態1にも適用可能である。また、上記実施の形態で説明した各構成要素を組み合わせて、新たな実施の形態とすることも可能である。
また、変換装置100の第1輝度変換(HPL→DPL)は例えば次の算式で変換する。
S2 = 2000/10000
M = 750/10000
a = 0.023
b = S1 - a*ln(S1) = 0.112105
つまり、第1輝度変換では、SDRの輝度値が、平均輝度値(CAL)と第1最大輝度値(CPL)との間である場合、自然対数を用いて、当該HDRの輝度値に対応するディスプレイ輝度値を決定する。
HDR映像のコンテンツピーク輝度やコンテンツ平均輝度等の情報を用いてHDR映像を変換することにより、コンテンツに応じて変換式を変えることができ、HDRの階調をなるべく保つように変換することが可能となる。また、暗すぎる、明るすぎるといった悪影響を抑制することができる。具体的には、HDR映像のコンテンツピーク輝度をディスプレイピーク輝度にマッピングすることにより、階調をなるべく保つようにしている。また、平均輝度付近以下の画素値を変えないことにより、全体的な明るさが変わらないようにしている。
以上、本開示の一つまたは複数の態様に係る再生方法および再生装置について、実施の形態に基づいて説明したが、本開示は、この実施の形態に限定されるものではない。本開示の趣旨を逸脱しない限り、当業者が思いつく各種変形を本実施の形態に施したもの、異なる実施の形態における構成要素を組み合わせて構築される形態なども、本開示の一つまたは複数の態様の範囲内に含まれてもよい。
101 EOTF変換部
102 輝度変換部
103 逆輝度変換部
104 逆SDRのEOTF変換部
200 表示装置
201 表示設定部
202 SDRのEOTF変換部
203 輝度変換部
204 表示部
400 データ出力装置
401 ビデオ復号部
402 外部メタ取得部
403 HDRメタ解釈部
404 HDR制御情報生成部
405 DR変換部
406 HDMI出力部
Claims (20)
- 映像ストリームを復号することで第1輝度範囲の第1映像信号を生成する復号部と、
前記第1映像信号の輝度範囲に関する第1メタデータを取得する取得部と、
前記第1メタデータを解釈することで、前記第1映像信号の輝度範囲を示す特性データを取得する解釈部と、
前記特性データを、所定の伝送プロトコルに従った制御情報に変換する制御情報生成部と、
前記制御情報を前記伝送プロトコルで出力する出力部とを備える
データ出力装置。 - 前記解釈部は、さらに、前記第1メタデータを解釈することで、前記第1映像信号の輝度範囲を変換するための変換補助データを取得し、
前記データ出力装置は、さらに、
前記変換補助データに基づいて、前記第1映像信号の輝度範囲を変換することで前記第1輝度範囲より狭い第2輝度範囲の第2映像信号を生成する変換部を備え、
前記出力部は、さらに、前記第1映像信号及び前記第2映像信号の少なくとも一方を前記伝送プロトコルで出力する
請求項1記載のデータ出力装置。 - 前記復号部は、さらに、前記映像ストリームから前記第1映像信号の輝度範囲に関する第2メタデータを取得し、
前記第1メタデータ又は前記第2メタデータは、映像信号の連続再生単位に含まれる複数の画像に対して共通に用いられ、前記特性データを含む静的メタデータを含み、
前記制御情報生成部は、前記静的メタデータに含まれる前記特性データを、前記伝送プロトコルに従った前記制御情報に変換し、
前記出力部は、前記第1映像信号を出力する場合、前記静的メタデータに基づく前記制御情報を出力する
請求項2記載のデータ出力装置。 - 前記第1メタデータ又は前記第2メタデータは、さらに、映像信号の連続再生単位より細かい単位に対して共通に用いられ、前記特性データを含む動的メタデータを含み、
前記制御情報生成部は、前記静的メタデータに含まれる前記特性データ及び前記動的メタデータに含まれる前記特性データを、前記伝送プロトコルに従った前記制御情報に変換し、
前記出力部は、前記第1映像信号を出力する場合、前記静的メタデータ及び前記動的メタデータに基づく前記制御情報を出力する
請求項3記載のデータ出力装置。 - 前記復号部は、さらに、前記映像ストリームから前記第1映像信号の輝度範囲に関する第2メタデータを取得し、
前記解釈部は、前記第1メタデータ及び前記第2メタデータの少なくとも一方を解析することで前記特性データ及び前記変換補助データを取得する
請求項2記載のデータ出力装置。 - 前記第1メタデータ及び前記第2メタデータは、前記第1映像信号の連続再生単位に含まれる複数の画像に対して共通に用いられる静的メタデータであり、
前記解釈部は、前記第1メタデータ及び前記第2メタデータを共に取得した場合、前記第1メタデータを解析することで前記特性データ及び前記変換補助データを取得する
請求項5記載のデータ出力装置。 - 前記第1メタデータ及び前記第2メタデータは、前記第1映像信号の連続再生単位に含まれる複数の画像に対して共通に用いられる静的メタデータであり、
前記解釈部は、
前記第1メタデータ及び前記第2メタデータのうち前記第1メタデータのみが取得された場合、前記第1メタデータを解析することで前記特性データ及び前記変換補助データを取得し、
前記第2メタデータが取得された場合、使用するメタデータを前記第1メタデータから前記第2メタデータに切り替える
請求項5記載のデータ出力装置。 - 前記出力部は、前記データ出力装置に接続されている表示装置が前記第1映像信号の輝度範囲の映像出力に対応している場合、前記第1映像信号及び前記制御情報を前記表示装置へ出力する
請求項2~7のいずれか1項に記載のデータ出力装置。 - 前記出力部は、前記表示装置が前記第1映像信号の輝度範囲の映像出力に対応していか否かを、前記伝送プロトコルの初期化処理において判定する
請求項8記載のデータ出力装置。 - 前記出力部は、前記第1メタデータが有効になる第2映像信号と同期して前記制御情報を出力する
請求項2~9のいずれか1項に記載のデータ出力装置。 - 前記第1メタデータは、当該第1メタデータの種別を示すタイプ情報を含み、
前記解釈部は、
前記タイプ情報を用いて、前記データ出力装置が当該第1メタデータを解釈可能であるかを判定し、
前記データ出力装置が当該第1メタデータを解釈可能である場合、当該第1メタデータを解釈することで前記特性データ及び前記変換補助データを取得する
請求項2~10のいずれか1項に記載のデータ出力装置。 - 前記第1メタデータは、HEVC(High Efficiency Video Coding)のSEI(Supplemental Enhancement Information)メッセージに格納されており、
前記SEIメッセージは、当該SEIメッセージにメタデータが含まれるか否かを示す識別情報を含み、
前記出力部は、
当該SEIメッセージにメタデータが含まれることを示す前記識別情報を含むSEIメッセージが取得され、かつ、データ出力先の表示装置が前記制御情報の入力に対応している場合には、前記SEIメッセージを前記伝送プロトコルに従ってそのまま出力する
請求項1~11のいずれか1項に記載のデータ出力装置。 - 前記出力部は、前記表示装置が前記第1映像信号の輝度範囲の映像出力に対応していない場合、前記第2映像信号及び前記制御情報を前記表示装置へ出力するとともに、前記第1輝度範囲から前記第2輝度範囲に変換された前記第2映像信号が表示される旨を前記表示装置に表示させる
請求項8記載のデータ出力装置。 - 前記出力部は、前記表示装置が前記第1映像信号の輝度範囲の映像出力に対応していない場合、前記第1輝度範囲から前記第2輝度範囲に変換された前記第2映像信号を表示させるかどうかを、ユーザが選択するためのメッセージを前記表示装置に表示させる
請求項8記載のデータ出力装置。 - 前記出力部は、前記表示装置が前記第1映像信号の輝度範囲の映像出力に対応していない場合、前記第1輝度範囲を前記第2輝度範囲に変換するための複数の変換方式のうちいずれかを、ユーザが選択するためのメッセージを前記表示装置に表示させる
請求項8記載のデータ出力装置。 - データ出力装置が行うデータ出力方法であって、
映像ストリームを復号することで第1輝度範囲の第1映像信号を生成する復号ステップと、
前記第1映像信号の輝度範囲に関する第1メタデータを取得する取得ステップと、
前記第1メタデータを解釈することで、前記第1映像信号の輝度範囲を示す特性データを取得する解釈ステップと、
前記特性データを、所定の伝送プロトコルに従った制御情報に変換する制御情報生成ステップと、
前記制御情報を前記伝送プロトコルで出力する出力ステップとを含む
データ出力方法。 - データ生成装置が行うデータ生成方法であって、
映像信号の輝度範囲に関するメタデータを生成する第1生成ステップと、
前記映像信号と前記メタデータとを含む映像ストリームを生成する第2生成ステップとを含み、
前記メタデータは、前記映像信号の輝度範囲を示す特性データを含む
データ生成方法。 - 前記メタデータは、解釈が必須である基本部と、解釈が任意である拡張部とを含む
請求項17記載のデータ生成方法。 - 前記メタデータは、ペイロードと、前記ペイロードのデータが基本部であるか拡張部であるかを示す情報と、前記ペイロードのデータサイズを示す情報とを含む
請求項18記載のデータ生成方法。 - 前記メタデータは、HEVC(High Efficiency Video Coding)のSEI(Supplemental Enhancement Information)メッセージに格納され、
前記SEIメッセージは、当該SEIメッセージにメタデータが含まれるか否かを示す識別情報を含む
請求項17~19のいずれか1項に記載のデータ生成方法。
Priority Applications (11)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
MX2016016602A MX358934B (es) | 2014-06-26 | 2015-06-15 | Dispositivo de salida de datos, metodo de salida de datos y metodo de generacion de datos. |
EP15811044.5A EP3163890B1 (en) | 2014-06-26 | 2015-06-15 | Data output device, data output method, and data generation method |
EP20174192.3A EP3713247A1 (en) | 2014-06-26 | 2015-06-15 | Data output device, data output method, and data generation method |
JP2015547182A JP5906504B1 (ja) | 2014-06-26 | 2015-06-15 | データ出力装置、データ出力方法及びデータ生成方法 |
CN201580019519.7A CN106165403B (zh) | 2014-06-26 | 2015-06-15 | 数据输出装置、数据输出方法以及数据生成方法 |
CN201911073813.2A CN110675840B (zh) | 2014-06-26 | 2015-06-15 | 显示装置及显示方法 |
MX2018010725A MX366608B (es) | 2014-06-26 | 2015-06-15 | Dispositivo de salida de datos, metodo de salida de datos y metodo de generacion de datos. |
US15/383,209 US10306175B2 (en) | 2014-06-26 | 2016-12-19 | Method for generating control information based on characteristic data included in metadata |
US15/596,218 US10291874B2 (en) | 2014-06-26 | 2017-05-16 | Method for generating control information based on characteristic data included in metadata |
US16/378,948 US10666891B2 (en) | 2014-06-26 | 2019-04-09 | Method for generating control information based on characteristic data included in metadata |
US16/850,314 US11140354B2 (en) | 2014-06-26 | 2020-04-16 | Method for generating control information based on characteristic data included in metadata |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201462017577P | 2014-06-26 | 2014-06-26 | |
US62/017,577 | 2014-06-26 | ||
JP2015-108904 | 2015-05-28 | ||
JP2015108904 | 2015-05-28 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/383,209 Continuation US10306175B2 (en) | 2014-06-26 | 2016-12-19 | Method for generating control information based on characteristic data included in metadata |
US15/383,209 Continuation-In-Part US10306175B2 (en) | 2014-06-26 | 2016-12-19 | Method for generating control information based on characteristic data included in metadata |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2015198553A1 true WO2015198553A1 (ja) | 2015-12-30 |
Family
ID=54937664
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2015/002973 WO2015198553A1 (ja) | 2014-06-26 | 2015-06-15 | データ出力装置、データ出力方法及びデータ生成方法 |
Country Status (6)
Country | Link |
---|---|
US (4) | US10306175B2 (ja) |
EP (2) | EP3713247A1 (ja) |
JP (4) | JP5906504B1 (ja) |
CN (3) | CN110708439A (ja) |
MX (1) | MX358934B (ja) |
WO (1) | WO2015198553A1 (ja) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPWO2016039172A1 (ja) * | 2014-09-12 | 2017-06-22 | ソニー株式会社 | 再生装置、再生方法、情報処理装置、情報処理方法、プログラム、および記録媒体 |
JP2017184249A (ja) * | 2014-06-26 | 2017-10-05 | パナソニックIpマネジメント株式会社 | 表示装置、表示方法及びコンピュータプログラム |
JP2018007194A (ja) * | 2016-07-08 | 2018-01-11 | 三菱電機株式会社 | 映像記録装置および映像再生装置 |
EP3293731A1 (en) * | 2016-09-12 | 2018-03-14 | Onkyo Corporation | Video processing device |
CN108702479A (zh) * | 2016-02-19 | 2018-10-23 | 索尼公司 | 信息处理设备、信息处理方法和程序 |
WO2018193715A1 (ja) * | 2017-04-21 | 2018-10-25 | パナソニックIpマネジメント株式会社 | 再生装置、再生方法、表示装置、および、表示方法 |
JP2020014246A (ja) * | 2019-10-15 | 2020-01-23 | ソニー株式会社 | 送信装置、送信方法、受信装置および受信方法 |
KR20200077513A (ko) * | 2017-10-31 | 2020-06-30 | 소니 주식회사 | 재생 장치, 재생 방법, 프로그램, 및 기록 매체 |
EP3603059A4 (en) * | 2017-03-21 | 2020-09-09 | Roku, Inc. | DISPLAY DIAGNOSTICS TO IMPROVE THE PERFORMANCE OF DISPLAY DEVICES |
JPWO2019098054A1 (ja) * | 2017-11-17 | 2020-11-26 | ソニー株式会社 | 情報処理装置、情報処理方法、記録媒体、再生装置、再生方法、およびプログラム |
WO2022230274A1 (ja) * | 2021-04-30 | 2022-11-03 | ソニーグループ株式会社 | 情報処理システム、および情報処理方法 |
Families Citing this family (42)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8606073B2 (en) * | 2010-05-12 | 2013-12-10 | Woodman Labs, Inc. | Broadcast management system |
EP3157242B1 (en) * | 2014-06-10 | 2019-07-03 | Panasonic Intellectual Property Management Co., Ltd. | Display system, display method, and display device |
JP6466258B2 (ja) * | 2014-08-07 | 2019-02-06 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生装置、再生方法および記録媒体 |
WO2016039025A1 (ja) * | 2014-09-08 | 2016-03-17 | ソニー株式会社 | 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム |
US10910017B2 (en) * | 2014-09-12 | 2021-02-02 | Sony Corporation | Reproducing device, reproducing method, information processing device, information processing method, program, and recording medium |
CN113613009B (zh) | 2015-02-27 | 2024-06-25 | 索尼公司 | 发送装置、发送方法、接收装置和接收方法 |
US10595099B2 (en) * | 2015-04-05 | 2020-03-17 | Lg Electronics Inc. | Method and device for transmitting and receiving broadcast signal for broadcast service on basis of XML subtitle |
US10516915B2 (en) * | 2015-04-24 | 2019-12-24 | Lg Electronics Inc. | Apparatus and method of processing metadata for indicating a change in HDR information |
US10122928B2 (en) * | 2015-09-09 | 2018-11-06 | Red.Com, Llc | Motion video output for multiple displays |
JP6535560B2 (ja) * | 2015-09-18 | 2019-06-26 | 東芝映像ソリューション株式会社 | 電子機器、及び表示方法 |
RU2728766C2 (ru) * | 2015-11-09 | 2020-07-31 | Томсон Лайсенсинг | Способ и устройство для адаптации видеоконтента, декодированного из элементарных потоков, под характеристики устройства отображения |
JP6765919B2 (ja) * | 2016-09-27 | 2020-10-07 | キヤノン株式会社 | 画像処理装置、その制御方法、プログラム |
KR20190088532A (ko) * | 2016-12-01 | 2019-07-26 | 엘지전자 주식회사 | 영상표시장치, 및 이를 구비하는 영상표시 시스템 |
JP6583584B2 (ja) | 2017-03-17 | 2019-10-02 | ヤマハ株式会社 | コンテンツ再生機器、コンテンツ再生方法、及びコンテンツ再生システム |
US9986200B1 (en) * | 2017-05-11 | 2018-05-29 | Novatek Microelectronics Corp. | Method and video conversion system of updating video setting |
CN109691119B (zh) * | 2017-06-21 | 2022-04-19 | 松下知识产权经营株式会社 | 影像显示***及影像显示方法 |
KR102351095B1 (ko) * | 2017-06-23 | 2022-01-13 | 삼성전자주식회사 | 전자 장치, 디스플레이 장치 및 그 제어 방법 |
US10939158B2 (en) * | 2017-06-23 | 2021-03-02 | Samsung Electronics Co., Ltd. | Electronic apparatus, display apparatus and control method thereof |
WO2019008818A1 (ja) * | 2017-07-07 | 2019-01-10 | パナソニックIpマネジメント株式会社 | 映像処理システム及び映像処理方法 |
KR102553764B1 (ko) * | 2017-07-13 | 2023-07-10 | 삼성전자주식회사 | 전자 장치, 디스플레이 장치 및 그 제어 방법 |
US10735688B2 (en) * | 2017-07-13 | 2020-08-04 | Samsung Electronics Co., Ltd. | Electronics apparatus, display apparatus and control method thereof |
CN109691118B (zh) | 2017-07-14 | 2021-12-17 | 松下知识产权经营株式会社 | 影像显示***及影像显示方法 |
JP2019028368A (ja) | 2017-08-02 | 2019-02-21 | 株式会社ソニー・インタラクティブエンタテインメント | レンダリング装置、ヘッドマウントディスプレイ、画像伝送方法、および画像補正方法 |
EP3684061B1 (en) | 2017-09-13 | 2021-07-07 | Panasonic Intellectual Property Management Co., Ltd. | Video display device and video display method |
JP7207314B2 (ja) * | 2017-09-21 | 2023-01-18 | ソニーグループ株式会社 | 再生装置、再生方法、プログラム |
US10992877B2 (en) * | 2017-10-06 | 2021-04-27 | Panasonic Intellectual Property Management Co., Ltd. | Video display apparatus and video display method |
US10972767B2 (en) * | 2017-11-01 | 2021-04-06 | Realtek Semiconductor Corp. | Device and method of handling multiple formats of a video sequence |
KR102448497B1 (ko) * | 2017-11-17 | 2022-09-28 | 삼성전자주식회사 | 디스플레이 장치, 디스플레이 장치의 제어 방법 및 영상 제공 장치 |
US10972674B2 (en) | 2017-12-27 | 2021-04-06 | Canon Kabushiki Kaisha | Electronic apparatus |
US10720124B2 (en) * | 2018-01-15 | 2020-07-21 | Microsoft Technology Licensing, Llc | Variable pixel rate display interfaces |
JP2019140430A (ja) * | 2018-02-06 | 2019-08-22 | 船井電機株式会社 | 再生装置 |
US10609424B2 (en) * | 2018-03-09 | 2020-03-31 | Dolby Laboratories Licensing Corporation | Single-layer progressive coding for supporting multi-capability HDR composition |
TW201946474A (zh) * | 2018-04-30 | 2019-12-01 | 圓剛科技股份有限公司 | 影像錄製設定方法 |
US10477141B1 (en) | 2018-05-17 | 2019-11-12 | Futurewei Technologies, Inc. | Frame synchronous packet switching for high-definition multimedia interface (HDMI) video transitions |
EP3588964A1 (en) * | 2018-06-26 | 2020-01-01 | InterDigital VC Holdings, Inc. | Metadata translation in hdr distribution |
CN108900926A (zh) * | 2018-06-29 | 2018-11-27 | 广州视源电子科技股份有限公司 | 一种电视预警方法、电视接收机与计算机可读存储介质 |
JP7344654B2 (ja) | 2019-03-07 | 2023-09-14 | キヤノン株式会社 | 撮像装置及び再生装置及びそれらの制御方法及びプログラム |
FR3101589B1 (fr) * | 2019-10-04 | 2023-04-14 | Valeo Systemes Dessuyage | Dispositif de nettoyage de vitre de véhicule |
CN112351264B (zh) * | 2020-11-04 | 2023-05-23 | 深圳Tcl数字技术有限公司 | 显示精度调整方法、装置、终端设备及存储介质 |
CN113870769B (zh) * | 2021-09-23 | 2023-03-21 | 深圳市艾比森光电股份有限公司 | 显示屏***的数据处理方法和*** |
US11606605B1 (en) | 2021-09-30 | 2023-03-14 | Samsung Electronics Co., Ltd. | Standard dynamic range (SDR) / hybrid log-gamma (HLG) with high dynamic range (HDR) 10+ |
CN114727029B (zh) * | 2022-03-21 | 2024-03-22 | 北京百度网讯科技有限公司 | 视频的处理方法、装置、电子设备及存储介质 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012147010A1 (en) * | 2011-04-28 | 2012-11-01 | Koninklijke Philips Electronics N.V. | Method and apparatus for generating an image coding signal |
WO2013046095A1 (en) * | 2011-09-27 | 2013-04-04 | Koninklijke Philips Electronics N.V. | Apparatus and method for dynamic range transforming of images |
WO2014002901A1 (ja) * | 2012-06-29 | 2014-01-03 | ソニー株式会社 | 画像処理装置および方法 |
Family Cites Families (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3322945B2 (ja) * | 1993-07-15 | 2002-09-09 | 株式会社デジタル | 表示制御装置 |
JP2003316314A (ja) * | 2001-12-27 | 2003-11-07 | Sony Corp | プラズマ表示装置、並びにその輝度補正方法およびその表示方法 |
CN100531340C (zh) * | 2004-04-12 | 2009-08-19 | 三菱电机株式会社 | 显示装置 |
WO2007037379A1 (ja) * | 2005-09-30 | 2007-04-05 | Matsushita Electric Industrial Co., Ltd. | 無線伝送システム |
US8194997B2 (en) | 2006-03-24 | 2012-06-05 | Sharp Laboratories Of America, Inc. | Methods and systems for tone mapping messaging |
KR20080015626A (ko) * | 2006-08-16 | 2008-02-20 | 삼성전자주식회사 | 영상처리장치 및 영상처리방법 |
JP2008054169A (ja) * | 2006-08-28 | 2008-03-06 | Orion Denki Kk | 外部表示装置に合わせた出力映像の解像度変更手段を備える映像再生装置 |
JP5145017B2 (ja) | 2006-12-05 | 2013-02-13 | 日本放送協会 | 画像信号処理装置 |
CN102354489B (zh) * | 2008-05-08 | 2014-01-22 | 香港应用科技研究院有限公司 | 增强图像动态范围的方法和装置 |
JP5304211B2 (ja) * | 2008-12-11 | 2013-10-02 | ソニー株式会社 | 表示装置、輝度調整装置、バックライト装置、輝度調整方法及びプログラム |
JP5566048B2 (ja) * | 2009-04-28 | 2014-08-06 | キヤノン株式会社 | 表示装置及びその制御方法 |
CN102055992B (zh) * | 2009-10-29 | 2012-06-27 | 华映视讯(吴江)有限公司 | 形成立体影像的方法及其相关显示模组 |
JP2011119828A (ja) * | 2009-12-01 | 2011-06-16 | Renesas Electronics Corp | 映像変換装置、記録媒体読取装置、映像再生装置及び映像再生情報送信方法 |
JP5899120B2 (ja) * | 2010-03-03 | 2016-04-06 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | カラーレジームを定義する装置及び方法 |
JP4861494B2 (ja) | 2010-05-31 | 2012-01-25 | 株式会社東芝 | 映像出力制御装置及び映像出力制御方法 |
EP2689392B1 (en) | 2011-03-24 | 2015-09-02 | Koninklijke Philips N.V. | Apparatus and method for analyzing image gradings |
US8743291B2 (en) * | 2011-04-12 | 2014-06-03 | Dolby Laboratories Licensing Corporation | Quality assessment for images that have extended dynamic ranges or wide color gamuts |
ES2899979T3 (es) * | 2011-05-10 | 2022-03-15 | Koninklijke Philips Nv | Generación y procesamiento de señales de imagen de rango dinámico alto |
RU2605435C2 (ru) * | 2011-06-14 | 2016-12-20 | Конинклейке Филипс Н.В. | Графическая обработка для видео с расширенным динамическим диапазоном |
US9451292B2 (en) * | 2011-09-15 | 2016-09-20 | Dolby Laboratories Licensing Corporation | Method and system for backward compatible, extended dynamic range encoding of video |
EP2579591A1 (en) * | 2011-10-04 | 2013-04-10 | Thomson Licensing | Method of and device for encoding an HDR image, method of and device for reconstructing an HDR image and non-transitory storage medium |
JP5002734B2 (ja) * | 2012-02-22 | 2012-08-15 | 株式会社東芝 | 映像出力制御装置及び映像出力制御方法 |
PL2880624T3 (pl) * | 2012-03-26 | 2019-12-31 | Koninklijke Philips N.V. | Oparte na obszarze jasności urządzenia i sposoby kodowania i dekodowania obrazu hdr |
WO2014043005A1 (en) * | 2012-09-12 | 2014-03-20 | Dolby Laboratories Licensing Corporation | Display management for images with enhanced dynamic range |
TWI618397B (zh) * | 2012-12-21 | 2018-03-11 | 杜比實驗室特許公司 | 在高位元深度視訊的可適性編碼中,高精度升取樣 |
ES2703517T3 (es) * | 2013-01-25 | 2019-03-11 | Dolby Laboratories Licensing Corp | Modulación de luz basada en la gestión de presentación global |
CN105009580B (zh) * | 2013-02-21 | 2017-09-29 | 杜比实验室特许公司 | 高动态范围视频的显示管理 |
EP3011741B1 (en) * | 2013-06-17 | 2018-03-07 | Dolby Laboratories Licensing Corporation | Adaptive reshaping for layered coding of enhanced dynamic range signals |
CN105531992A (zh) * | 2013-09-06 | 2016-04-27 | Lg电子株式会社 | 用于在数字广播***中发送和接收用于高动态范围呈现的超高清晰度广播信号的方法和装置 |
WO2015076608A1 (ko) * | 2013-11-21 | 2015-05-28 | 엘지전자 주식회사 | 비디오 처리 방법 및 비디오 처리 장치 |
CN110708439A (zh) | 2014-06-26 | 2020-01-17 | 松下知识产权经营株式会社 | 显示装置及数据输出方法 |
US20160034504A1 (en) * | 2014-07-31 | 2016-02-04 | AppDynamics, Inc. | Efficient aggregation, storage and querying of large volume metrics |
-
2015
- 2015-06-15 CN CN201911073815.1A patent/CN110708439A/zh active Pending
- 2015-06-15 MX MX2016016602A patent/MX358934B/es active IP Right Grant
- 2015-06-15 JP JP2015547182A patent/JP5906504B1/ja active Active
- 2015-06-15 WO PCT/JP2015/002973 patent/WO2015198553A1/ja active Application Filing
- 2015-06-15 EP EP20174192.3A patent/EP3713247A1/en not_active Withdrawn
- 2015-06-15 EP EP15811044.5A patent/EP3163890B1/en active Active
- 2015-06-15 CN CN201911073813.2A patent/CN110675840B/zh active Active
- 2015-06-15 CN CN201580019519.7A patent/CN106165403B/zh active Active
-
2016
- 2016-03-04 JP JP2016041674A patent/JP5995032B1/ja active Active
- 2016-08-09 JP JP2016156588A patent/JP6569912B2/ja active Active
- 2016-12-19 US US15/383,209 patent/US10306175B2/en active Active
-
2017
- 2017-05-09 JP JP2017093419A patent/JP6671022B2/ja active Active
- 2017-05-16 US US15/596,218 patent/US10291874B2/en active Active
-
2019
- 2019-04-09 US US16/378,948 patent/US10666891B2/en active Active
-
2020
- 2020-04-16 US US16/850,314 patent/US11140354B2/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012147010A1 (en) * | 2011-04-28 | 2012-11-01 | Koninklijke Philips Electronics N.V. | Method and apparatus for generating an image coding signal |
WO2013046095A1 (en) * | 2011-09-27 | 2013-04-04 | Koninklijke Philips Electronics N.V. | Apparatus and method for dynamic range transforming of images |
WO2014002901A1 (ja) * | 2012-06-29 | 2014-01-03 | ソニー株式会社 | 画像処理装置および方法 |
Cited By (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10666891B2 (en) | 2014-06-26 | 2020-05-26 | Panasonic Intellectual Property Management Co., Ltd. | Method for generating control information based on characteristic data included in metadata |
JP2017184249A (ja) * | 2014-06-26 | 2017-10-05 | パナソニックIpマネジメント株式会社 | 表示装置、表示方法及びコンピュータプログラム |
US10306175B2 (en) | 2014-06-26 | 2019-05-28 | Panasonic Intellectual Property Management Co., Ltd. | Method for generating control information based on characteristic data included in metadata |
US10291874B2 (en) | 2014-06-26 | 2019-05-14 | Panasonic Intellectual Property Management Co., Ltd. | Method for generating control information based on characteristic data included in metadata |
US11140354B2 (en) | 2014-06-26 | 2021-10-05 | Panasonic Intellectual Property Management Co., Ltd. | Method for generating control information based on characteristic data included in metadata |
JPWO2016039172A1 (ja) * | 2014-09-12 | 2017-06-22 | ソニー株式会社 | 再生装置、再生方法、情報処理装置、情報処理方法、プログラム、および記録媒体 |
US10965926B2 (en) | 2014-09-12 | 2021-03-30 | Sony Corporation | Reproducing apparatus, reproducing method, information processing apparatus, information processing method, program, and recording medium |
CN108702479A (zh) * | 2016-02-19 | 2018-10-23 | 索尼公司 | 信息处理设备、信息处理方法和程序 |
EP3419281A4 (en) * | 2016-02-19 | 2018-12-26 | Sony Corporation | Image processing device, image processing method, and program |
JP2018007194A (ja) * | 2016-07-08 | 2018-01-11 | 三菱電機株式会社 | 映像記録装置および映像再生装置 |
US10250862B2 (en) | 2016-09-12 | 2019-04-02 | Onkyo Corporation | Video processing device |
JP2018046317A (ja) * | 2016-09-12 | 2018-03-22 | オンキヨー株式会社 | 映像処理装置 |
EP3293731A1 (en) * | 2016-09-12 | 2018-03-14 | Onkyo Corporation | Video processing device |
EP3603059A4 (en) * | 2017-03-21 | 2020-09-09 | Roku, Inc. | DISPLAY DIAGNOSTICS TO IMPROVE THE PERFORMANCE OF DISPLAY DEVICES |
WO2018193715A1 (ja) * | 2017-04-21 | 2018-10-25 | パナソニックIpマネジメント株式会社 | 再生装置、再生方法、表示装置、および、表示方法 |
JPWO2018193715A1 (ja) * | 2017-04-21 | 2020-02-27 | パナソニックIpマネジメント株式会社 | 再生装置、再生方法、表示装置、および、表示方法 |
CN109644289A (zh) * | 2017-04-21 | 2019-04-16 | 松下知识产权经营株式会社 | 再现装置、再现方法、显示装置、以及显示方法 |
CN109644289B (zh) * | 2017-04-21 | 2021-11-09 | 松下知识产权经营株式会社 | 再现装置、再现方法、显示装置、以及显示方法 |
JP7018567B2 (ja) | 2017-04-21 | 2022-02-14 | パナソニックIpマネジメント株式会社 | 再生装置、再生方法、表示装置、および、表示方法 |
KR20200077513A (ko) * | 2017-10-31 | 2020-06-30 | 소니 주식회사 | 재생 장치, 재생 방법, 프로그램, 및 기록 매체 |
KR102558213B1 (ko) * | 2017-10-31 | 2023-07-24 | 소니그룹주식회사 | 재생 장치, 재생 방법, 프로그램, 및 기록 매체 |
JPWO2019098054A1 (ja) * | 2017-11-17 | 2020-11-26 | ソニー株式会社 | 情報処理装置、情報処理方法、記録媒体、再生装置、再生方法、およびプログラム |
JP7180610B2 (ja) | 2017-11-17 | 2022-11-30 | ソニーグループ株式会社 | 再生装置、再生方法、およびプログラム |
JP7435704B2 (ja) | 2017-11-17 | 2024-02-21 | ソニーグループ株式会社 | 再生装置、再生方法、プログラム、情報処理システム |
JP2020014246A (ja) * | 2019-10-15 | 2020-01-23 | ソニー株式会社 | 送信装置、送信方法、受信装置および受信方法 |
WO2022230274A1 (ja) * | 2021-04-30 | 2022-11-03 | ソニーグループ株式会社 | 情報処理システム、および情報処理方法 |
Also Published As
Publication number | Publication date |
---|---|
JP2016208499A (ja) | 2016-12-08 |
CN106165403B (zh) | 2019-11-29 |
JP5906504B1 (ja) | 2016-04-20 |
EP3163890B1 (en) | 2020-08-05 |
US20170105042A1 (en) | 2017-04-13 |
JP2017022717A (ja) | 2017-01-26 |
CN110675840A (zh) | 2020-01-10 |
EP3713247A1 (en) | 2020-09-23 |
JP2017184249A (ja) | 2017-10-05 |
US20190238788A1 (en) | 2019-08-01 |
US20170251161A1 (en) | 2017-08-31 |
MX358934B (es) | 2018-09-10 |
US20200244916A1 (en) | 2020-07-30 |
EP3163890A1 (en) | 2017-05-03 |
CN106165403A (zh) | 2016-11-23 |
US10291874B2 (en) | 2019-05-14 |
US10666891B2 (en) | 2020-05-26 |
US10306175B2 (en) | 2019-05-28 |
JPWO2015198553A1 (ja) | 2017-04-20 |
JP6671022B2 (ja) | 2020-03-25 |
JP6569912B2 (ja) | 2019-09-04 |
US11140354B2 (en) | 2021-10-05 |
CN110708439A (zh) | 2020-01-17 |
EP3163890A4 (en) | 2017-05-03 |
JP5995032B1 (ja) | 2016-09-21 |
CN110675840B (zh) | 2022-03-25 |
MX2016016602A (es) | 2017-04-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6569912B2 (ja) | 表示装置及びデータ出力方法 | |
JP6566320B2 (ja) | データ出力装置及びデータ出力方法 | |
JP6573238B2 (ja) | 表示装置、変換装置、表示方法、および、コンピュータプログラム | |
JP6566271B2 (ja) | 伝送方法及び再生装置 | |
JP6512458B2 (ja) | データ再生方法及び再生装置 | |
WO2015198552A1 (ja) | コンテンツデータ生成方法、映像ストリーム伝送方法及び映像表示方法 | |
JP6731627B2 (ja) | 再生装置で実行される方法、及び再生装置 | |
JP2017139511A (ja) | コンテンツデータ生成方法、映像ストリーム伝送方法及び映像表示方法 | |
JP7170236B2 (ja) | 再生装置 | |
JP6928885B2 (ja) | 表示装置、表示方法及びコンピュータプログラム | |
JP6643669B2 (ja) | 表示装置および表示方法 | |
JP6751893B2 (ja) | 再生方法、再生装置、表示方法及び表示装置 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2015547182 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15811044 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2016/016602 Country of ref document: MX |
|
REEP | Request for entry into the european phase |
Ref document number: 2015811044 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2015811044 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |