WO2018077876A1 - Method for adapting a luminance of a multimedia content, corresponding computer program product and apparatus - Google Patents

Method for adapting a luminance of a multimedia content, corresponding computer program product and apparatus Download PDF

Info

Publication number
WO2018077876A1
WO2018077876A1 PCT/EP2017/077161 EP2017077161W WO2018077876A1 WO 2018077876 A1 WO2018077876 A1 WO 2018077876A1 EP 2017077161 W EP2017077161 W EP 2017077161W WO 2018077876 A1 WO2018077876 A1 WO 2018077876A1
Authority
WO
WIPO (PCT)
Prior art keywords
luminance
visual
loudness
image
pixels
Prior art date
Application number
PCT/EP2017/077161
Other languages
French (fr)
Inventor
Erik Reinhard
Mozhdeh Seifi
Mikael LE PENDU
Fatma HAWARY
Original Assignee
Thomson Licensing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing filed Critical Thomson Licensing
Publication of WO2018077876A1 publication Critical patent/WO2018077876A1/en

Links

Classifications

    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G5/00Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
    • G09G5/10Intensity circuits
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2320/00Control of display operating conditions
    • G09G2320/06Adjustment of display parameters
    • G09G2320/0626Adjustment of display parameters for control of overall brightness
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2320/00Control of display operating conditions
    • G09G2320/06Adjustment of display parameters
    • G09G2320/0626Adjustment of display parameters for control of overall brightness
    • G09G2320/0646Modulation of illumination source brightness and image signal correlated to each other
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2320/00Control of display operating conditions
    • G09G2320/06Adjustment of display parameters
    • G09G2320/0686Adjustment of display parameters with two or more screen areas displaying information with different brightness or colours
    • GPHYSICS
    • G09EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
    • G09GARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
    • G09G2360/00Aspects of the architecture of display systems
    • G09G2360/16Calculation or use of calculated indices related to luminance levels in display data

Definitions

  • the field of the disclosure is that of the image and video processing.
  • the disclosure relates to a method for adapting the luminance of a video content to a reference so as to normalize the rendering.
  • the disclosure can be of interest in any field where the rendering of different contents originating from different sources is involved. This can be the case for instance in the field of broadcasting where channels need to support contents from various sources, with a range of formats and characteristics.
  • Broadcasting channels typically take contents from various sources, with a range of formats and characteristics. This includes for instance movies, live content, television shows and commercials. Here, aside from various format conversions, it is important that the look and feel between programs is consistent for a pleasant user experience.
  • HDR high dynamic range
  • post-production services typically produce two different grades, one for cinema viewing in dark environments, and one for home viewing in dim environments. There is reasonable uniformity in the set-up of the mastering suites across grading studios. However, at least one third grade may be introduced to serve HDR cinema and/or high dynamic range home viewing;
  • high dynamic range display devices incorporate power management strategies to limit the power output. In practice this means that the highest achievable luminance on such displays can only be achieved for a certain proportion of the screen, possibly for only a certain amount of time.
  • a particular aspect of the present disclosure relates to a method for adapting a luminance of a multimedia content.
  • Such method comprises, for a current pixel of an image of the multimedia content: • obtaining a visual loudness value from at least a luminance value of the current pixel and from at least a luminance value of another pixel of the image;
  • the visual loudness being a measure representative of a luminance of the image with respect to a spatial configuration of distinct brightness areas in the image.
  • the present disclosure proposes a new and inventive solution for adapting the luminance of a multimedia content (called a source content) that includes image or video materials to a reference (a reference image for example), for example to ensure consistency between two contents to be broadcast.
  • a source content a multimedia content
  • a reference a reference image for example
  • a visual loudness metric representative of the perception of how light or dark an image is seen by a person viewing the content is introduced. It indeed appears that the lightness of a patch of color in an image depends on the luminance of that patch in relation to other areas of the image. It is thus proposed to introduce a visual loudness metric of the images in the content corresponding to a measure representative of the luminance of the image with respect to the spatial configuration of distinct brightness areas in it.
  • a luminance of a pixel in the image can be adapted by attributing a reference luminance value to this pixel, this reference luminance value being associated to the visual loudness value of the pixel.
  • this reference luminance value being associated to the visual loudness value of the pixel.
  • the "global" luminance of the source image can be adapted so as to get an adapted image that presents a visual consistency with the selected reference.
  • Another aspect of the present disclosure relates to an apparatus for adapting a luminance of a multimedia content.
  • Such apparatus comprises a memory and a processor configured for, for a current image of the multimedia content:
  • the visual loudness being a measure representative of a luminance of the image with respect to a spatial configuration of distinct brightness areas in the image.
  • the obtaining a visual loudness value comprises:
  • V h a highlight visual loudness component V h , representing an assessment of the contribution of highlight pixels in a luminance of the image with respect to a spatial configuration of areas containing the highlight pixels in the image;
  • V m a mid-tone visual loudness component V m , representing an assessment of the contribution of mid-tone pixels in a luminance of the image with respect to a spatial configuration of areas containing the mid-tone pixels in the image
  • V d a dark visual loudness component V d , representing an assessment of the contribution of dark pixels in a luminance of the image with respect to a spatial configuration of areas containing the dark pixels in the image;
  • the three groups of pixels i.e. highlight, mid-tone and dark pixels contribute to the sensation of visual loudness in different ways. Taking into account for their respective contribution separately allows achieving a suitable visual consistency between contents (for example a source content and a reference content).
  • the reference luminance value being associated to the visual loudness value is obtained from a luminance value of a pixel in a reference image.
  • the luminance of the image in the multimedia content can be adapted so as to present a visual consistency with a reference image, e.g. in a reference content.
  • the method further comprises, or the apparatus is further configured for, averaging the reference luminance value with the luminance value of the current pixel, providing a modified luminance value, the adapting the luminance of the multimedia content corresponding to attributing the modified luminance value to the current pixel.
  • the obtaining at least two visual loudness components comprises calculating at least one threshold belonging to the group comprising:
  • the highlight pixels corresponding to pixels with a luminance above the upper threshold T mh the dark pixels corresponding to pixels with a luminance below the lower threshold T dm and the mid-tone pixels corresponding to pixels with a luminance below the upper threshold T mh and above the lower threshold T dm .
  • the three groups of pixels are defined in a simple and robust way.
  • the obtaining at least two visual loudness components comprises deriving at least two corresponding characteristic luminance values belonging to the group comprising:
  • the upper threshold T mh being a function of the characteristic highlight luminance value L h and of the characteristic mid-tone luminance value L m
  • the lower threshold T dm being a function of the characteristic mid-tone luminance value L m and of the characteristic dark luminance value L d .
  • the characteristic luminance of the three groups of pixels is defined in a simple and robust way through such characteristic luminance value.
  • the characteristic highlight luminance value L h corresponds to a first percentage, respectively second percentage, of pixels in the image having a luminance below the characteristic highlight luminance value L h , respectively characteristic dark luminance value L d , the first percentage being greater than the second percentage.
  • the characteristic mid-tones luminance value L m corresponds to a geometric mean luminance of pixels in the image.
  • the characteristic highlights, mid-tones and darks luminance values are derived based on robust statistics, i.e. on statistics that are not sensitive to outliers (such as minimum and maximum values in an image).
  • the visual loudness value is extrapolated from the at least two visual loudness components and from the at least two corresponding characteristic luminance values.
  • a visual loudness value can be derived in a simple and robust way for each pixel in the considered image, based only on the visual loudness components associated to the three groups of pixels and on the corresponding characteristic luminance values.
  • the obtaining at least two visual loudness components comprises:
  • the highlight visual loudness component V h being a function of a density of highlight pixels of the at least one sub-mask and of a size of the at least one sub-mask.
  • areas of high density of lighter pixels can be obtained in order to derive a highlight visual loudness component that takes into account for the spatial configuration of those areas.
  • the function is a non-linear function of the size of the at least one sub-mask.
  • the splitting is enforced hierarchically and recursively on previously obtained sub-masks according to a quad-tree structure.
  • the mid-tone visual loudness component V m is a function of a barycenter of a luminance of the mid-tone pixels with respect to the characteristic highlight L h and/or dark L d luminance values.
  • the dark visual loudness component V d is a function of a barycenter of a luminance of the dark pixels with respect to the characteristic dark luminance value L d and/or of a zero luminance reference value.
  • the mid-tone and dark visual loudness components can be derived in a simple and straightforward manner from the characteristic highlight and dark luminance values.
  • the function of a barycenter is a non-linear function.
  • the contribution of the mid-tone and dark visual loudness components to the overall visual loudness of the image can be controlled more accurately.
  • the multimedia content is a video content
  • the obtaining a visual loudness value further comprises adjusting visual loudness components for providing at least two adjusted visual loudness components as a function of the at least two visual loudness components and of at least two previous visual loudness components obtained for a previous image in the video content
  • the visual loudness value being obtained from the at least two adjusted visual loudness components.
  • the disclosed method is temporally stable so that no flicker or other unwanted artefacts are introduced in the video content.
  • the method according to this embodiment allows a smooth adaptation of the content so as to follow the speed of human visual adaptation for instance.
  • Another aspect of the present disclosure relates to a computer program product comprising program code instructions for implementing the above-mentioned method for adapting a luminance of a multimedia content (in any of its different embodiments), when the program is executed on a computer or a processor.
  • Another aspect of the present disclosure relates to a non-transitory computer-readable carrier medium storing a computer program product which, when executed by a computer or a processor causes the computer or the processor to carry out the above-mentioned method for adapting a luminance of a multimedia content (in any of its different embodiments).
  • FIGS. 1 a and 1 b are flowcharts of particular embodiments of the disclosed method for adapting a luminance of a multimedia content
  • Figure 2 is a schematic illustration of the structural blocks of an exemplary apparatus that can be used for implementing the method for adapting a luminance of a multimedia content according to any of the embodiments disclosed in relation with figures 1 a and 1 b.
  • the general principle of the disclosed method consists in obtaining a visual loudness value for a multimedia content, the visual loudness value representing a measure of how light or dark an image appears with respect to a spatial configuration of distinct brightness areas in the image, in order to adjust the multimedia content for matching the visual loudness of a reference multimedia content.
  • the disclosed method calculates a visual loudness value from at least a luminance value of a current pixel in an image of the multimedia content (including image or video materials) and of at least a luminance value of another pixel in the same image. Then, the disclosed method associates a reference luminance value to the obtained visual loudness value, and adaptes the luminance value of the multimedia content by attributing the reference luminance value to the current pixel.
  • FIG. 1 a and 1 b we present different embodiments of a method for adapting a luminance of a multimedia content.
  • a visual loudness value is obtained, for a multimedia content, from at least a luminance of a current pixel in an image of the multimedia content and of at least a luminance value of another pixel in the same image.
  • a visual loudness representative of how light or dark an image appears with respect to a spatial configuration of distinct brightness areas in the image, is expressed as a function of components among three visual loudness components belonging to the group comprising: • a highlight visual loudness component V h , representing an assessment of the contribution of highlight pixels in a luminance of the image with respect to a spatial configuration of areas containing the highlight pixels in the image;
  • V m a mid-tone visual loudness component
  • a dark visual loudness component V d representing an assessment of the contribution of dark pixels in a luminance of the image with respect to a spatial configuration of areas containing the dark pixels in the image.
  • lightness is a perceptual measure, which can be defined as the attribute of visual sensation according to which the area in which the visual stimulus is presented appears to emit more or less light in proportion to that emitted by a similarly illuminated area perceived as the "white” stimulus (see “Reinhard, E., Khan, E. A., Akyuz, A. O., & Johnson, G. (2008). Color imaging: fundamentals and applications , CRC Press”).
  • computational lightness models may be used, ideally but not necessarily including a measure of size.
  • computational measures of lightness evaluate high luminance pixels, as opposed to medium and low luminance pixels
  • the overall impression of a frame or image is in part determined by the luminance level of the mid-tones, in relation to the dark and light parts of the frame or image.
  • a mid-tones visual loudness component can thus be assessed by deriving measures from the analysis of the histogram of an image.
  • the highlight visual loudness component V h component is following modern insights into the perception of lightness to assess the contribution of the lightest parts of the image.
  • the mid-tones visual loudness component V m constitutes an analysis of the histogram of the (possibly down- sampled) image to evaluate the contribution of mid-tones.
  • the darks visual loudness component V d results from an analysis of the spatial extent of the darkest parts of the scene.
  • the overall measure V of visual loudness of the considered image is the product of the three contributions:
  • the visual loudness may be represented by pairs of numbers, e.g. three pairs of numbers characterizing the visual loudness contributions of the dark, mid-tone and highlight regions in the considered picture. In that case, these pairs of numbers are
  • L h is a characteristic highlight luminance value representative of a luminance of lighter pixels in the image
  • L m is a characteristic mid-tone luminance value representative of an average luminance of pixels in the image.
  • L d is a characteristic dark luminance value representative of a luminance of darker pixels in the image.
  • a curve mapping any luminance value to a visual loudness contribution can be obtained, so that a notion of visual loudness can be assigned to each luminance value in the image.
  • a visual loudness value is extrapolated from the at least two visual loudness components and from the at least two corresponding characteristic luminance values and can be assigned to each luminance value in the image.
  • a visual loudness value can thus be derived in a simple and robust way for each pixel in the considered image, based only on the visual loudness components associated to the three groups of pixels and on the corresponding characteristic luminance values.
  • the characteristic highlight luminance value L h the characteristic midtone luminance value L m and the characteristic dark luminance value L d are derived.
  • the characteristic mid-tones luminance value L m corresponds to a geometric mean luminance of pixels in the image.
  • the geometric mean luminance L m is a measure of where the majority of the luminance of pixels typically is located within the dynamic range of the image and can be calculated as follows:
  • L(x, y) is the luminance of the pixel located at position (x, y) in the image.
  • this is the arithmetic mean luminance that is used for deriving
  • the characteristic highlight luminance value L h respectively characteristic dark luminance value L d
  • the characteristic dark luminance value L d may be characterized by setting the value of N to 0%, 1 %, 5% or 10%, or any other appropriate value that suitably separates dark luminance values from all other luminance values.
  • the characteristic highlight luminance value L h may be characterized by setting the value of N to a high value, such as 90%, 95%, 99% or 100%.
  • the characteristic highlight, mid-tone and dark luminance values are derived based on robust statistics, i.e. on statistics that are not sensitive to outliers (such as minimum and maximum values in an image).
  • At least one threshold belonging to the group comprising an upper threshold T mh and a lower threshold T dm is calculated.
  • those upper and lower thresholds allow defining the highlight pixels as the pixels having a luminance above the upper threshold T mh , the dark pixels as the pixels with a luminance below the lower threshold T dm , and the mid-tone pixels as the pixels with a luminance below the upper threshold T mh and above the lower threshold T dm .
  • the three groups of pixels are defined in a simple and robust way.
  • the upper threshold T mh is calculated as a function of the characteristic highlight luminance value L h and of the characteristic mid-tone luminance value L m ;
  • the lower threshold T dm is calculated as a function of the characteristic mid-tone luminance value L m and of the characteristic dark luminance value L d .
  • the upper threshold T mh between mid-tone pixels and highlight pixels may be computed analogously following:
  • the threshold T dm is computed by finding the minimum value of the image's luminance histogram between L d and L m .
  • the threshold T mh can be computed by finding the minimum value of the image's luminance histogram between L m and L h .
  • the highlight visual loudness component V h is obtained as a function of all the pixels with a luminance value larger than T mh , i.e.:
  • V h is a function f h Q of the set of pixels S h that contains all pixels that have a luminance larger than the threshold T mh .
  • the /c-means clustering algorithm (or one of its heuristic variants) could be used to partition an image into k clusters.
  • This algorithm is also known as Lloyd's algorithm.
  • Other approaches include the use of Gaussian Mixture models that are trained with the Expectation-Maximization algorithm, /c-medians clustering, Fuzzy C-Means Clustering and /c-means++.
  • the number of clusters k needs to be known a priori. As in the present case, this number is not known, it could be estimated with a separate algorithm, or using a rule of thumb. Further, these methods do not provide information regarding the size of each cluster.
  • Suitable algorithms that do not require the number of clusters to be known a priori include Density-Based Spatial Clustering of Applications with Noise (DBSCAN), Ordering Points To Identify the Clustering Structure (OPTICS), and hierarchical clustering such as SLINK (for "single-link” clustering method) and CLINK (for "complete link” clustering method).
  • DBSCAN Density-Based Spatial Clustering of Applications with Noise
  • OTICS Ordering Points To Identify the Clustering Structure
  • hierarchical clustering such as SLINK (for "single-link” clustering method) and CLINK (for "complete link” clustering method).
  • a mask M h representative of a location of the highlight pixels in the considered image is determined in block 100a3 (figure 1 b).
  • the mask M h is created with the same size as the image, with values set to "0" or "1 " dependent on whether the corresponding pixels belong to the set S h or not, i.e.:
  • the mask M h is split when a density of highlight pixels in the mask M h is below a highlight pixels density threshold t for providing at least one sub-mask presenting a higher density of highlight pixels than the mask M h .
  • the mask M h is split into four smaller masks, each of size w/2 by h/2 if the number of pixels corresponding to elements with a value of "1 " in the mask is lower than the highlight pixels density threshold t, i.e. if the following inequalities hold true:
  • the highlight pixels density threshold t is suitably chosen in the range t tow , t high .
  • An example value of t !ow may be 0.2, whereas a typical example value of thigh may be 0.8 for achieving good results at a reasonable computation cost.
  • the mask M h is split either iteratively or recursively into four equal parts, provided the splitting criterion is satisfied.
  • the splitting is enforced hierarchically and recursively on previously obtained sub-masks according to a quad-tree structure.
  • thresholds on the width and height of the (sub-)masks ensure that the depth of the quad-tree stays within reasonable bounds.
  • the resulting quad-tree stores at every node the calculated percentage p.
  • Each node q is located at a given level in the quad-tree and represents a sub-mask.
  • the size s of a leaf node i.e. the size of the corresponding sub-mask
  • the height of the quad-tree i.e. the maximum level
  • each leaf node q has a size s t and a percentage of pixels p; that correspond to a value of "1 " in the mask M h .
  • the size s depends on the level I at which this node is located as discussed above.
  • hQ maps the size of each leaf node to a scaling factor that is applied to pj .
  • the function hQ is a non-linear function of the size s; of the sub-mask represented by the leaf node q t . This allows taking into account possible power management issues that could arise in certain display devices. This function then models the inability of a display to reach certain luminance levels for regions of a certain size.
  • the highlight visual loudness component V h induced by the highlight pixels is obtained as the sum of each region's contribution normalized by the sum of the region's sizes s t involved in the calculation, i.e.:
  • the calculation of the mid-tone visual loudness component V m is obtained as a function of all the pixels with a luminance value larger than T dm , but lower than
  • the luminance histogram of images provides clues as to whether the scene was overall light, medium or dark, irrespective of whether the luminance values of the image signify absolute or relative values.
  • the luminance histogram of a high dynamic range image can be interpreted as a probability density function.
  • high dynamic range image histograms tend to deviate from a Gaussian distribution in that they often exhibit high values for their third and fourth moments (skew and kurtosis).
  • the kurtosis represents the fact that the distribution is long-tailed. It would therefore be possible to analyze the histogram for skew and kurtosis, and make decisions as to whether the image is overall light or dark.
  • the mid-tone visual loudness component V m is obtained in a more straightforward manner using the characteristic luminance values.
  • the mid-tone visual loudness component V m is obtained as a function of a ratio r expressed in a f s:
  • the ratio r may be defined relative to the characteristic highlight luminance value L h instead of the characteristic dark luminance value L d , i.e.:
  • the ratio r may be defined as a function of a ratio r expressed in a form of a barycenter relative to both the characteristic highlight luminance value
  • the mid-tone visual loudness component V m is obtained as a function of r, i.e. as a function of a barycenter of a luminance of the mid-tone pixels with respect to the characteristic highlight L h and/or dark L d luminance values, i.e. in a very simple and straightforward manner from those characteristic values.
  • the ratio r is defined relative of the upper threshold T mh and/or of the lower threshold T dm instead of characteristic highlight L h and/or dark L d luminance values.
  • the three characteristic luminance values L d , L m and L h are used to assess the contribution of the mid-tones to visual loudness by computing the ratio r
  • the ratio r which may be computed according to any of the embodiment disclosed above, may be normalized.
  • the normalization of r is achieved as follows:
  • the ratio r norm may be weighted according to a non-linear function, which may be any function that takes a normalized scalar as input and produces a scalar as output. This includes, but is not limited to power functions, logarithmic functions, exponential functions, trigonometric and sigmoidal functions.
  • a non-linear function which may be any function that takes a normalized scalar as input and produces a scalar as output. This includes, but is not limited to power functions, logarithmic functions, exponential functions, trigonometric and sigmoidal functions.
  • the advantage of such weighting is that the contribution to overall visual loudness of the mid-tone component can be better controlled.
  • the ratio r norm weighted according to such non-linear function may be directly interpreted as the mid-tone visual loudness component V m .
  • the dark visual loudness component V d is obtained as a function of a ratio r d expressed in a form of a barycenter referred to a zero luminance reference value
  • the ratio r d is expressed in a form of a barycenter referred to the characteristic dark luminance value L d instead of the zero luminance reference value.
  • r d may be normalized to another quantity than the lower threshold Tdm (e.g. the characteristic dark luminance value L d ).
  • the dark visual loudness component V d is thus obtained as a function of a barycenter of the luminance of the dark pixels with respect to the characteristic dark luminance value L d and/or of a zero luminance reference value, i.e. in a very simple and straightforward manner from those characteristic values.
  • r d may be weighted according to a non-linear function, which may be any function that takes a normalized scalar as input and produces a scalar as output.
  • the advantage of such weighting is that the contribution to overall visual loudness of the dark component can be better controlled. In that case, the ratio r d weighted according to such non-linear function may be directly interpreted as the dark visual loudness component V d .
  • the multimedia content includes a video content
  • the at least two visual loudness components obtained in block 100a are further adjusted for providing at least two adjusted visual loudness components as a function of the at least two visual loudness components and of at least two previous visual loudness components obtained for a previous image in the video content.
  • a leaky integration is applied to the visual loudness components V d , V m and V h and/or to the characteristic luminance values L d , L m and L h obtained for successive images in the video content.
  • the disclosed method is temporally stable so that no flicker or other unwanted artefacts are introduced in the video content.
  • the method according to this embodiment allows a smooth adaptation of the content so as to follow the speed of human visual adaptation for instance.
  • these parameters may be adjusted such that the overall visual loudness V follows a temporal Cornsweet profile (see “Cornsweet, T. (2012). “Visual perception”. Academic press”). This could be used to increase the visual appearance of temporal contrast, or alternatively to decrease the visual appearance of temporal contrast.
  • the visual loudness value to be obtained from at least the luminance of the current pixel i.e. the output of block 100
  • the visual loudness value to be obtained from at least the luminance of the current pixel is calculated as a function of the at least two visual loudness components obtained in block 100a.
  • the multimedia content includes a video content
  • the visual loudness value to be obtained is calculated in block 100c as a function of the at least two adjusted visual loudness components obtained in block 100b.
  • the visual loudness value is thus obtained as extrapolated from the at least two visual loudness components (or two adjusted visual loudness components) and from the at least two corresponding characteristic luminance values, based on the luminance L image (x, y) of the current pixel located at position (x, y).
  • a reference luminance value is associated to the visual loudness value obtained in block 100.
  • the reference luminance values are stored in a look-up table with entries corresponding to all possible visual loudness values.
  • a reference source of content may be created that has a desirable level of visual loudness.
  • the visual loudness for this reference content is represented by at least two among the three pairs of characteristic luminance values / visual loudness components:
  • the function c ref should be monotonically increasing. It is envisaged that the desired visual loudness for an HDR broadcast channel is encoded with this function, and should be applied to all program material to be broadcast over this channel. It is therefore anticipated that it is possible to construct a reference visual loudness presentation that, if represented by a curve, yields a monotonically increasing curve. This reference curve must be monotonically increasing so that it can be inverted.
  • the reference luminance value being associated to the visual loudness value associated to the current pixel is obtained from a luminance value of a pixel in a reference image.
  • the luminance of the considered image in the multimedia content can be adapted so as to present a visual consistency with a reference image in a reference source of content.
  • the reference luminance value L ref (L image (x, y)) is averaged with the luminance value L image (x, y) of the current pixel in the considered image, providing a modified luminance value.
  • the luminance of the multimedia content is adapted by attributing the reference luminance value obtained in block 1 10 to the current pixel.
  • the "global" luminance of the source image can be adapted so as to get an adapted image that presents a visual consistency with the selected reference.
  • the reference luminance value L ref (L image (x, y)) is averaged with the luminance value L image (x, y) of the current pixel in the considered image for providing a modified luminance value in block 130, this is the modified luminance value L' ref (L image (x, y)) that is attributed to the current pixel.
  • L' ref L image (x, y)
  • an apparatus 200 for implementing the disclosed method comprises a non-volatile memory 201 (e.g. a read-only memory (ROM) or a hard disk), a volatile memory 203 (e.g. a random access memory or RAM) and a processor 202.
  • the non-volatile memory 201 is a non-transitory computer-readable carrier medium.
  • Non-volatile memory 201 and volatile memory 203 may be combined into a single entity referred to as memory, configured as combined volatile and non-volatile memory, or just one of the types of memory.
  • the aforementioned program code instructions are transferred from the non-volatile memory 201 to the volatile memory 203 so as to be executed by the processor 202.
  • the volatile memory 203 likewise includes registers for storing the variables and parameters required for this execution.
  • This program code instructions can be stored in a non- transitory computer-readable carrier medium that is detachable (for example a floppy disk, a CD-ROM or a DVD-ROM) or non-detachable; or
  • a dedicated machine or component such as an FPGA (Field Programmable Gate Array), an ASIC (Application-Specific Integrated Circuit) or any dedicated hardware component.
  • FPGA Field Programmable Gate Array
  • ASIC Application-Specific Integrated Circuit
  • the disclosure is not limited to a purely software-based implementation, in the form of computer program instructions, but that it may also be implemented in hardware form or any form combining a hardware portion and a software portion.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Hardware Design (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)

Abstract

A method is proposed for adapting a luminance of a multimedia content. Such method comprises, for a current pixel of an image of the multimedia content: - obtaining (100) a visual loudness value from at least a luminance value of the current pixel and from at least a luminance value of another pixel of the image; - associating (110) a reference luminance value to the visual loudness value; and - adapting (120) the luminance of the multimedia content by attributing the reference luminance value to the current pixel; the visual loudness being a measure representative of a luminance of the image with respect to a spatial configuration of distinct brightness areas in the image.

Description

Method for adapting a luminance of a multimedia content, corresponding computer program product and apparatus
1. FIELD OF THE DISCLOSURE
The field of the disclosure is that of the image and video processing.
More specifically, the disclosure relates to a method for adapting the luminance of a video content to a reference so as to normalize the rendering.
The disclosure can be of interest in any field where the rendering of different contents originating from different sources is involved. This can be the case for instance in the field of broadcasting where channels need to support contents from various sources, with a range of formats and characteristics.
2. TECHNOLOGICAL BACKGROUND
This section is intended to introduce the reader to various aspects of art, which may be related to various aspects of the present disclosure that are described and/or claimed below. This discussion is believed to be helpful in providing the reader with background information to facilitate a better understanding of the various aspects of the present disclosure. Accordingly, it should be understood that these statements are to be read in this light, and not as admissions of prior art.
Broadcasting channels typically take contents from various sources, with a range of formats and characteristics. This includes for instance movies, live content, television shows and commercials. Here, aside from various format conversions, it is important that the look and feel between programs is consistent for a pleasant user experience.
For that, broadcasters typically strive to maintain visual consistency of the content that they transmit. This means that luminance differences between the various different types of content need to be reconciled, while maintaining the director's intent of each piece of content. There are many tone reproduction and inverse tone reproduction techniques to adjust the luminance profile of content.
However, the emergence of high dynamic range (HDR) imaging may limit the results obtained with existing technics. Indeed, various developments are observed related to HDR: · whereas conventional standard dynamic range contents have a well-defined peak luminance, assumed to be 100 cd/m2 (nits), the diversification of camera equipment and grading practices means that contents will have different peak luminances. Further adjustments to video content may be necessary to make content consistent with the look and feel of the channel. Here, a balance needs to be achieved between channel consistency and maintaining the director's intent of the original content;
• currently, post-production services typically produce two different grades, one for cinema viewing in dark environments, and one for home viewing in dim environments. There is reasonable uniformity in the set-up of the mastering suites across grading studios. However, at least one third grade may be introduced to serve HDR cinema and/or high dynamic range home viewing;
• advertisers are likely to use the headroom afforded by HDR imaging systems in ways that make their content stand out. In essence, it is anticipated that advertisements will regularly appear to be significantly brighter unless it is controlled in a manner that balances the needs of the broadcasting channel, the viewers, and the advertisers;
• to achieve high luminance levels, high dynamic range display devices incorporate power management strategies to limit the power output. In practice this means that the highest achievable luminance on such displays can only be achieved for a certain proportion of the screen, possibly for only a certain amount of time.
Currently, algorithms exist that prepare content either for displays with a higher dynamic range than the content, or for displays with a lower dynamic range than the content. Often, the dynamic range of the content is matched to the dynamic range of the display, with the aid of knowledge of the dynamic range extrema. In practice this means that the black level and the peak luminance (or only the peak luminance) of both content and display are used to guide the dynamic range adjustment.
However, despite its advantages (e.g. the relative pixel values are maintained, which means that all contrasts in an image can be enhanced or reduced in substantially similar ways) it appears that such level adjustments are not sufficient if the look and feel of the content needs to be brought closer together, or needs to be matched.
There is thus a need for a method that allows adapting the look and feel of an image or a video content for matching a reference.
There is also a need for this method to be temporally stable in order to avoid introducing any flicker or artifacts when applied to a video content.
3. SUMMARY
A particular aspect of the present disclosure relates to a method for adapting a luminance of a multimedia content. Such method comprises, for a current pixel of an image of the multimedia content: • obtaining a visual loudness value from at least a luminance value of the current pixel and from at least a luminance value of another pixel of the image;
• associating a reference luminance value to the visual loudness value; and
• adapting the luminance of the multimedia content by attributing the reference luminance value to the current pixel;
the visual loudness being a measure representative of a luminance of the image with respect to a spatial configuration of distinct brightness areas in the image.
Thus, the present disclosure proposes a new and inventive solution for adapting the luminance of a multimedia content (called a source content) that includes image or video materials to a reference (a reference image for example), for example to ensure consistency between two contents to be broadcast.
For this to be possible, a visual loudness metric representative of the perception of how light or dark an image is seen by a person viewing the content is introduced. It indeed appears that the lightness of a patch of color in an image depends on the luminance of that patch in relation to other areas of the image. It is thus proposed to introduce a visual loudness metric of the images in the content corresponding to a measure representative of the luminance of the image with respect to the spatial configuration of distinct brightness areas in it.
Based on such metric, a luminance of a pixel in the image can be adapted by attributing a reference luminance value to this pixel, this reference luminance value being associated to the visual loudness value of the pixel. In other words, the "global" luminance of the source image can be adapted so as to get an adapted image that presents a visual consistency with the selected reference.
Another aspect of the present disclosure relates to an apparatus for adapting a luminance of a multimedia content. Such apparatus comprises a memory and a processor configured for, for a current image of the multimedia content:
• obtaining a visual loudness value from at least a luminance value of the current pixel and from at least a luminance value of another pixel of the image;
• associating a reference luminance value to the visual loudness value; and
• adapting the luminance of the multimedia content by attributing the reference luminance value to the current pixel;
the visual loudness being a measure representative of a luminance of the image with respect to a spatial configuration of distinct brightness areas in the image. Such an apparatus is particularly adapted for implementing the method for adapting a luminance of a multimedia content according to the present disclosure. Thus, the characteristics and advantages of this apparatus are the same as the disclosed method for adapting a luminance of a multimedia content.
According to one embodiment, the obtaining a visual loudness value comprises:
• obtaining at least two visual loudness components belonging to the group comprising:
o a highlight visual loudness component Vh, representing an assessment of the contribution of highlight pixels in a luminance of the image with respect to a spatial configuration of areas containing the highlight pixels in the image;
o a mid-tone visual loudness component Vm, representing an assessment of the contribution of mid-tone pixels in a luminance of the image with respect to a spatial configuration of areas containing the mid-tone pixels in the image; and
o a dark visual loudness component Vd, representing an assessment of the contribution of dark pixels in a luminance of the image with respect to a spatial configuration of areas containing the dark pixels in the image;
• calculating the visual loudness value as a function of the at least two visual loudness components.
Indeed, the three groups of pixels (i.e. highlight, mid-tone and dark pixels) contribute to the sensation of visual loudness in different ways. Taking into account for their respective contribution separately allows achieving a suitable visual consistency between contents (for example a source content and a reference content).
According to one embodiment, the reference luminance value being associated to the visual loudness value is obtained from a luminance value of a pixel in a reference image.
Thus, the luminance of the image in the multimedia content can be adapted so as to present a visual consistency with a reference image, e.g. in a reference content.
According to one embodiment, the method further comprises, or the apparatus is further configured for, averaging the reference luminance value with the luminance value of the current pixel, providing a modified luminance value, the adapting the luminance of the multimedia content corresponding to attributing the modified luminance value to the current pixel.
Thus, only a partial adaptation of the luminance of the content is performed so that a smooth adaptation of the source content is obtained. According to one embodiment, the obtaining at least two visual loudness components comprises calculating at least one threshold belonging to the group comprising:
• an upper threshold Tmh; and
• a lower threshold Tdm;
the highlight pixels corresponding to pixels with a luminance above the upper threshold Tmh, the dark pixels corresponding to pixels with a luminance below the lower threshold Tdm and the mid-tone pixels corresponding to pixels with a luminance below the upper threshold Tmh and above the lower threshold Tdm.
Thus, the three groups of pixels are defined in a simple and robust way.
According to one embodiment, the obtaining at least two visual loudness components comprises deriving at least two corresponding characteristic luminance values belonging to the group comprising:
• a characteristic highlight luminance value Lh representative of a luminance of lighter pixels in the image;
· a characteristic mid-tone luminance value Lm representative of an average luminance of pixels in the image; and
• a characteristic dark luminance value Ld representative of a luminance of darker pixels in the image;
the upper threshold Tmh being a function of the characteristic highlight luminance value Lh and of the characteristic mid-tone luminance value Lm, and the lower threshold Tdm being a function of the characteristic mid-tone luminance value Lm and of the characteristic dark luminance value Ld.
Thus, the characteristic luminance of the three groups of pixels is defined in a simple and robust way through such characteristic luminance value.
According to one embodiment, the characteristic highlight luminance value Lh, respectively characteristic dark luminance value Ld, corresponds to a first percentage, respectively second percentage, of pixels in the image having a luminance below the characteristic highlight luminance value Lh, respectively characteristic dark luminance value Ld, the first percentage being greater than the second percentage.
According to one embodiment, the characteristic mid-tones luminance value Lm corresponds to a geometric mean luminance of pixels in the image. Thus, the characteristic highlights, mid-tones and darks luminance values are derived based on robust statistics, i.e. on statistics that are not sensitive to outliers (such as minimum and maximum values in an image).
According to one embodiment, the visual loudness value is extrapolated from the at least two visual loudness components and from the at least two corresponding characteristic luminance values.
Thus, a visual loudness value can be derived in a simple and robust way for each pixel in the considered image, based only on the visual loudness components associated to the three groups of pixels and on the corresponding characteristic luminance values.
According to one embodiment, the obtaining at least two visual loudness components comprises:
• determining a mask Mh representative of a location of the highlight pixels in the image; and
• splitting the mask Mh when a density of the highlight pixels in the mask Mh is below a highlight pixels density threshold, the splitting providing at least one sub-mask presenting a higher density of highlight pixels than the mask Mh;
the highlight visual loudness component Vh being a function of a density of highlight pixels of the at least one sub-mask and of a size of the at least one sub-mask.
Thus, areas of high density of lighter pixels can be obtained in order to derive a highlight visual loudness component that takes into account for the spatial configuration of those areas.
According to one embodiment, the function is a non-linear function of the size of the at least one sub-mask.
Thus, the inability of certain HDR displays to reach certain luminance levels for regions of a certain size can be taken into account in the determination of the highlight visual loudness component.
According to one embodiment, the splitting is enforced hierarchically and recursively on previously obtained sub-masks according to a quad-tree structure.
According to one embodiment, the mid-tone visual loudness component Vm is a function of a barycenter of a luminance of the mid-tone pixels with respect to the characteristic highlight Lh and/or dark Ld luminance values. According to another embodiment, the dark visual loudness component Vd is a function of a barycenter of a luminance of the dark pixels with respect to the characteristic dark luminance value Ld and/or of a zero luminance reference value.
Thus, the mid-tone and dark visual loudness components can be derived in a simple and straightforward manner from the characteristic highlight and dark luminance values.
According to one embodiment, the function of a barycenter is a non-linear function.
Thus, the contribution of the mid-tone and dark visual loudness components to the overall visual loudness of the image can be controlled more accurately.
According to one embodiment, the multimedia content is a video content, and the obtaining a visual loudness value further comprises adjusting visual loudness components for providing at least two adjusted visual loudness components as a function of the at least two visual loudness components and of at least two previous visual loudness components obtained for a previous image in the video content,
the visual loudness value being obtained from the at least two adjusted visual loudness components.
Thus, the disclosed method is temporally stable so that no flicker or other unwanted artefacts are introduced in the video content.
Furthermore, when a sharp transition occurs in the content, or during the time immediately following such a sharp transition (where for instance a transition may be due to switching between a movie and a commercial and back), the method according to this embodiment allows a smooth adaptation of the content so as to follow the speed of human visual adaptation for instance.
Another aspect of the present disclosure relates to a computer program product comprising program code instructions for implementing the above-mentioned method for adapting a luminance of a multimedia content (in any of its different embodiments), when the program is executed on a computer or a processor.
Another aspect of the present disclosure relates to a non-transitory computer-readable carrier medium storing a computer program product which, when executed by a computer or a processor causes the computer or the processor to carry out the above-mentioned method for adapting a luminance of a multimedia content (in any of its different embodiments). 4. LIST OF FIGURES
Other features and advantages of embodiments shall appear from the following description, given by way of indicative and non-exhaustive examples and from the appended drawings, of which:
- Figures 1 a and 1 b are flowcharts of particular embodiments of the disclosed method for adapting a luminance of a multimedia content;
Figure 2 is a schematic illustration of the structural blocks of an exemplary apparatus that can be used for implementing the method for adapting a luminance of a multimedia content according to any of the embodiments disclosed in relation with figures 1 a and 1 b.
5. DETAILED DESCRIPTION
In all of the figures of the present document, the same numerical reference signs designate similar elements and steps.
The general principle of the disclosed method consists in obtaining a visual loudness value for a multimedia content, the visual loudness value representing a measure of how light or dark an image appears with respect to a spatial configuration of distinct brightness areas in the image, in order to adjust the multimedia content for matching the visual loudness of a reference multimedia content.
For that, the disclosed method calculates a visual loudness value from at least a luminance value of a current pixel in an image of the multimedia content (including image or video materials) and of at least a luminance value of another pixel in the same image. Then, the disclosed method associates a reference luminance value to the obtained visual loudness value, and adaptes the luminance value of the multimedia content by attributing the reference luminance value to the current pixel.
Referring now to figures 1 a and 1 b, we present different embodiments of a method for adapting a luminance of a multimedia content.
In block 100, a visual loudness value is obtained, for a multimedia content, from at least a luminance of a current pixel in an image of the multimedia content and of at least a luminance value of another pixel in the same image.
According to the present disclosure, a visual loudness, representative of how light or dark an image appears with respect to a spatial configuration of distinct brightness areas in the image, is expressed as a function of components among three visual loudness components belonging to the group comprising: • a highlight visual loudness component Vh, representing an assessment of the contribution of highlight pixels in a luminance of the image with respect to a spatial configuration of areas containing the highlight pixels in the image;
• a mid-tone visual loudness component Vm, representing an assessment of the contribution of mid-tone pixels in a luminance of the image with respect to a spatial configuration of areas containing the mid-tone pixels in the image; and
• a dark visual loudness component Vd, representing an assessment of the contribution of dark pixels in a luminance of the image with respect to a spatial configuration of areas containing the dark pixels in the image.
The motivation for this comes from the observation that dark pixels, mid-tone pixels and highlight pixels each contribute to the sensation of visual loudness in different ways.
Indeed, it appears that lightness is a perceptual measure, which can be defined as the attribute of visual sensation according to which the area in which the visual stimulus is presented appears to emit more or less light in proportion to that emitted by a similarly illuminated area perceived as the "white" stimulus (see "Reinhard, E., Khan, E. A., Akyuz, A. O., & Johnson, G. (2008). Color imaging: fundamentals and applications , CRC Press").
It was recently found that the lightness of a uniform patch of color depends on the luminance of that patch in relation to other regions, as well as its size (see "Allred, S. R., Radonjic, A., Gilchrist, A. L., & Brainard, D. H. (2012). "Lightness perception in high dynamic range images: Local and remote luminance effects ". Journal of vision, 12(2), 7') . This is known as the "area rule", formalized by Alan Gilchrist as: "In a simple display, when the darker of the two regions has the greater relative area, as the darker region grows in area, its lightness value goes up in direct proportion. At the same time the lighter region first appears white, then a fluorescent white and finally, self-luminous." (see "Gilchrist, A., Kossyfidis, C, Bonato, F., Agostini, T., Cataliotti, J., Li, X., ... & Economou, E. ( 1999). "An anchoring theory of lightness perception ". Psychological review, 106(4), 795').
In the definition of visual loudness, a variety of computational lightness models may be used, ideally but not necessarily including a measure of size. However, within the context of visual loudness, computational measures of lightness evaluate high luminance pixels, as opposed to medium and low luminance pixels
It moreover appears that the overall impression of a frame or image is in part determined by the luminance level of the mid-tones, in relation to the dark and light parts of the frame or image. A mid-tones visual loudness component can thus be assessed by deriving measures from the analysis of the histogram of an image.
Last, it also appears that the extent of the dark parts of the frame or image also impacts the impression of visual loudness. However, the magnitude of visual loudness will be inversely proportional to the percentage of pixels that can be considered dark.
Consequently, in one embodiment of the present disclosure, the highlight visual loudness component Vh component is following modern insights into the perception of lightness to assess the contribution of the lightest parts of the image. The mid-tones visual loudness component Vm constitutes an analysis of the histogram of the (possibly down- sampled) image to evaluate the contribution of mid-tones. The darks visual loudness component Vd results from an analysis of the spatial extent of the darkest parts of the scene.
In one exemplary embodiment, the overall measure V of visual loudness of the considered image is the product of the three contributions:
v = vh vm vd
In other embodiments, a visual loudness calculation is calculated that omits one or two of the components. This can be formalized by setting Vh = 1, Vm = 1 and/or Vd = 1 in the above equation.
In yet another embodiment, the visual loudness may be represented by pairs of numbers, e.g. three pairs of numbers characterizing the visual loudness contributions of the dark, mid-tone and highlight regions in the considered picture. In that case, these pairs of numbers are
(Lm, Vm) (Eq-1 )
M
where:
• Lh is a characteristic highlight luminance value representative of a luminance of lighter pixels in the image;
• Lm is a characteristic mid-tone luminance value representative of an average luminance of pixels in the image; and
• Ld is a characteristic dark luminance value representative of a luminance of darker pixels in the image.
Based on at least two of those three pairs of numbers, or control points, a curve mapping any luminance value to a visual loudness contribution can be obtained, so that a notion of visual loudness can be assigned to each luminance value in the image. In other words, a visual loudness value is extrapolated from the at least two visual loudness components and from the at least two corresponding characteristic luminance values and can be assigned to each luminance value in the image. A visual loudness value can thus be derived in a simple and robust way for each pixel in the considered image, based only on the visual loudness components associated to the three groups of pixels and on the corresponding characteristic luminance values.
There are many possible ways to define curves according to control points, including but not limited to splines such as B-splines or non-uniform rational basis splines (NURBS), piecewise linear curve segments and polynomials. Further, such curves can be represented for instance with 1 D Look-Up Tables (LUTs).
We now disclose in relation with block 100a (figure 1 b) example derivations for obtaining at least two visual loudness components belonging to the group comprising a highlight visual loudness component Vh, a mid-tone visual loudness component Vm, and a dark visual loudness component Vd.
For that, in block 100a1 (figure 1 b), the characteristic highlight luminance value Lh, the characteristic midtone luminance value Lm and the characteristic dark luminance value Ld are derived.
In one embodiment, the characteristic mid-tones luminance value Lm corresponds to a geometric mean luminance of pixels in the image. The geometric mean luminance Lm is a measure of where the majority of the luminance of pixels typically is located within the dynamic range of the image and can be calculated as follows:
Figure imgf000012_0001
where w and h are the width and the height of the image and L(x, y) is the luminance of the pixel located at position (x, y) in the image.
In another embodiment, this is the arithmetic mean luminance that is used for deriving
Lm, yet leading to simpler calculations.
Conversely, the characteristic highlight luminance value Lh, and the characteristic dark luminance value Ld, may be computed as a percentile, i.e. the luminance value for which N% of the pixels are below that luminance value. In other words, the characteristic highlight luminance value Lh, respectively characteristic dark luminance value Ld, corresponds to a first percentage, respectively second percentage, of pixels in the image having a luminance below the characteristic highlight luminance value Lh, respectively characteristic dark luminance value Ld, the first percentage being greater than the second percentage.
For instance, the characteristic dark luminance value Ld may be characterized by setting the value of N to 0%, 1 %, 5% or 10%, or any other appropriate value that suitably separates dark luminance values from all other luminance values. In the same way, the characteristic highlight luminance value Lh may be characterized by setting the value of N to a high value, such as 90%, 95%, 99% or 100%.
Consequently, the characteristic highlight, mid-tone and dark luminance values are derived based on robust statistics, i.e. on statistics that are not sensitive to outliers (such as minimum and maximum values in an image).
In block 100a2 (figure 1 b), at least one threshold belonging to the group comprising an upper threshold Tmh and a lower threshold Tdm is calculated.
More precisely, those upper and lower thresholds allow defining the highlight pixels as the pixels having a luminance above the upper threshold Tmh, the dark pixels as the pixels with a luminance below the lower threshold Tdm, and the mid-tone pixels as the pixels with a luminance below the upper threshold Tmh and above the lower threshold Tdm.
With such definition, the three groups of pixels are defined in a simple and robust way.
Having that the three characteristic luminance values Ld, Lm and Lh (introduced above in relation with equation (Eq-1 )) are representative of luminance values located in dark, mid, and high luminance regions, in one embodiment:
• the upper threshold Tmh is calculated as a function of the characteristic highlight luminance value Lh and of the characteristic mid-tone luminance value Lm; and
• the lower threshold Tdm is calculated as a function of the characteristic mid-tone luminance value Lm and of the characteristic dark luminance value Ld.
For example, the lower threshold Tdm between dark pixels and mid-tone pixels may be computed as an average between Ld and Lm according to: where a interpolates between the two luminance values. For a value of a = 0.5, for instance, the average between Ld and Lm is computed. The upper threshold Tmh between mid-tone pixels and highlight pixels may be computed analogously following:
Tmh = /?Lm + (l - /?)Lft
where β interpolates between the two luminance values Lm and Lh as does a for Ld and Lm. In another example, the threshold Tdm is computed by finding the minimum value of the image's luminance histogram between Ld and Lm. The threshold Tmh can be computed by finding the minimum value of the image's luminance histogram between Lm and Lh.
Back to block 100a (figure 1 b), the highlight visual loudness component Vh is obtained as a function of all the pixels with a luminance value larger than Tmh, i.e.:
Vh = fh Sh , where Sh = {Lt \ Lt > Tmh]
In other words, Vh is a function fhQ of the set of pixels Sh that contains all pixels that have a luminance larger than the threshold Tmh.
In that perspective, it is proposed to follow recent insights into lightness perception. In models of lightness perception, uniform patches with a given luminance are assessed with respect to both size and level. In particular, larger patches with a given luminance are perceived to be less light relative to smaller patches with the same luminance. This means that to be able to assess lightness, it is necessary to assess the pixels in the set Sh with respect to how they are grouped. In essence, if the highlight pixels in an image form many small clusters, then their perception will be different than if these highlight pixels form a small collection of much larger clusters.
There are several approaches possible to determine clusters for a collection of pixels. In exemplary embodiments, the /c-means clustering algorithm (or one of its heuristic variants) could be used to partition an image into k clusters. This algorithm is also known as Lloyd's algorithm. Other approaches include the use of Gaussian Mixture models that are trained with the Expectation-Maximization algorithm, /c-medians clustering, Fuzzy C-Means Clustering and /c-means++. In these algorithms the number of clusters k needs to be known a priori. As in the present case, this number is not known, it could be estimated with a separate algorithm, or using a rule of thumb. Further, these methods do not provide information regarding the size of each cluster.
Suitable algorithms that do not require the number of clusters to be known a priori include Density-Based Spatial Clustering of Applications with Noise (DBSCAN), Ordering Points To Identify the Clustering Structure (OPTICS), and hierarchical clustering such as SLINK (for "single-link" clustering method) and CLINK (for "complete link" clustering method).
As the specific application of highlight assessment does not require high accuracy in the detection of clusters, but does require knowledge of the size of such cluster, a different and straightforward approach is described which may serve as an exemplary implementation. For that, a mask Mh representative of a location of the highlight pixels in the considered image is determined in block 100a3 (figure 1 b).
In an exemplary embodiment, the mask Mh is created with the same size as the image, with values set to "0" or "1 " dependent on whether the corresponding pixels belong to the set Sh or not, i.e.:
h y y) lo otherwise
In block 100a4 (figure 1 b), the mask Mh is split when a density of highlight pixels in the mask Mh is below a highlight pixels density threshold t for providing at least one sub-mask presenting a higher density of highlight pixels than the mask Mh.
For instance, denoting respectively w and h the width and the height of the image, the mask Mh is split into four smaller masks, each of size w/2 by h/2 if the number of pixels corresponding to elements with a value of "1 " in the mask is lower than the highlight pixels density threshold t, i.e. if the following inequalities hold true:
w∑n Mh(x, y)
p = < t
w h
In one embodiment, the highlight pixels density threshold t is suitably chosen in the range ttow, thigh . An example value of t!ow may be 0.2, whereas a typical example value of thigh may be 0.8 for achieving good results at a reasonable computation cost.
In another embodiment, the mask Mh is split either iteratively or recursively into four equal parts, provided the splitting criterion is satisfied. In other words, the splitting is enforced hierarchically and recursively on previously obtained sub-masks according to a quad-tree structure.
In this embodiment, thresholds on the width and height of the (sub-)masks ensure that the depth of the quad-tree stays within reasonable bounds. Example values for these thresholds may be tw = th = 8.
The resulting quad-tree stores at every node the calculated percentage p. Each node q is located at a given level in the quad-tree and represents a sub-mask. The size s of a leaf node (i.e. the size of the corresponding sub-mask) relates directly to the level I of that node in the quad-tree, according to s = (w h)4~l (assuming that the root node is at level I = 0). The height of the quad-tree (i.e. the maximum level) is limited by the two thresholds on the width and height, tw and th respectively, discussed above.
Considering that the maximum level in the quad-tree is given by lmax, the calculation then proceeds by analyzing all leaf nodes located at levels 0 < I < lmax. These leaf nodes were not subdivided further, because the stopping criterion p ≥ t was reached. In other words, these leaf nodes contain a substantial number of highlight pixels. The set of leaf nodes between levels 0 < I < lmax is denoted Q = {qt, - , qn], where n is the number of leaf nodes in this set. Each leaf node q has a size st and a percentage of pixels p; that correspond to a value of "1 " in the mask Mh. The size s; depends on the level I at which this node is located as discussed above.
The visual loudness vi of each node qt is determined as follows:
vi = pi h(Si)
where the function hQ maps the size of each leaf node to a scaling factor that is applied to pj . This function is typically monotonically increasing. In its simplest form, it is the identity function, h(Si) = Si , but other functions can be considered.
In one exemplary embodiment, the function hQ is a non-linear function of the size s; of the sub-mask represented by the leaf node qt. This allows taking into account possible power management issues that could arise in certain display devices. This function then models the inability of a display to reach certain luminance levels for regions of a certain size.
Furthermore, following the findings in lightness perception disclosed in "Gilchrist, A., Kossyfidis, C, Bonato, F., Agostini, T., Cataliotti, J., Li, X., ... & Economou, E. (1999). "An anchoring theory of lightness perception ". Psychological review, 106(4), 795', it appears that the visual loudness increases as the size of the region st increases. It also appears that the visual loudness increases with the number of highlight pixels in each region s¾ .
Consequently, in an exemplary embodiment of the present disclosure, the highlight visual loudness component Vh induced by the highlight pixels is obtained as the sum of each region's contribution normalized by the sum of the region's sizes st involved in the calculation, i.e.:
Figure imgf000016_0001
Back to block 100a, the calculation of the mid-tone visual loudness component Vm is obtained as a function of all the pixels with a luminance value larger than Tdm , but lower than
V-m = fm Sm)> where Sm = {Li I Tdm < L; < Tmh}
In general, the luminance histogram of images provides clues as to whether the scene was overall light, medium or dark, irrespective of whether the luminance values of the image signify absolute or relative values. The luminance histogram of a high dynamic range image can be interpreted as a probability density function. In that case, high dynamic range image histograms tend to deviate from a Gaussian distribution in that they often exhibit high values for their third and fourth moments (skew and kurtosis). In particular, the kurtosis represents the fact that the distribution is long-tailed. It would therefore be possible to analyze the histogram for skew and kurtosis, and make decisions as to whether the image is overall light or dark.
However, in particular embodiments of the present disclosure, the mid-tone visual loudness component Vm is obtained in a more straightforward manner using the characteristic luminance values.
In one embodiment, the mid-tone visual loudness component Vm is obtained as a function of a ratio r expressed in a f s:
Figure imgf000017_0001
Small values of r indicate that the mid-tones of the image are relatively close to the dark values. This indicates that the image may be dark overall. On the other hand, large values of r indicate the reverse, i.e. that the image may be light overall.
In another embodiment, the ratio r may be defined relative to the characteristic highlight luminance value Lh instead of the characteristic dark luminance value Ld, i.e.:
Figure imgf000017_0002
Small values of r then indicate that the mid-tones of the image are relatively close to the highlight values. This indicates that the image may be light overall. On the other hand, large values of r indicate the reverse, i.e. that the image may be dark overall.
In yet another embodiment, the ratio r may be defined as a function of a ratio r expressed in a form of a barycenter relative to both the characteristic highlight luminance value
Lh and the characteristic dark luminance value Ld.
Whatever the considered embodiment discussed above, the mid-tone visual loudness component Vm is obtained as a function of r, i.e. as a function of a barycenter of a luminance of the mid-tone pixels with respect to the characteristic highlight Lh and/or dark Ld luminance values, i.e. in a very simple and straightforward manner from those characteristic values.
In other embodiments, the ratio r is defined relative of the upper threshold Tmh and/or of the lower threshold Tdm instead of characteristic highlight Lh and/or dark Ld luminance values. In another embodiment, the three characteristic luminance values Ld , Lm and Lh are used to assess the contribution of the mid-tones to visual loudness by computing the ratio r
Figure imgf000018_0001
In yet another embodiment, the ratio r, which may be computed according to any of the embodiment disclosed above, may be normalized. The normalization may be necessary as pixels with a luminance L; in the mid-tone luminance region are defined to have values in the range [Tdrn, Tmh] , with Ld < Tdm and Lh > Tmh. This means that without normalization r ranges between nlow = Tdm~Ld > o and nhiah = Tmh~Ld < i. The normalization of r is achieved as follows:
_ f — niow
'norm
nhigh nlow
Further, the ratio rnorm may be weighted according to a non-linear function, which may be any function that takes a normalized scalar as input and produces a scalar as output. This includes, but is not limited to power functions, logarithmic functions, exponential functions, trigonometric and sigmoidal functions. The advantage of such weighting is that the contribution to overall visual loudness of the mid-tone component can be better controlled. In that case, the ratio rnorm weighted according to such non-linear function may be directly interpreted as the mid-tone visual loudness component Vm .
In the same way, the calculation of the dark visual loudness component Vd is obtained as a function of all the pixels with a luminance value lower than Tdm , i.e.:
Vd = fd {Sd), where Sd = {L; | Lt < Tdm]
In one embodiment, the dark visual loudness component Vd is obtained as a function of a ratio rd expressed in a form of a barycenter referred to a zero luminance reference value
In alternative embodiments, the ratio rd is expressed in a form of a barycenter referred to the characteristic dark luminance value Ld instead of the zero luminance reference value.
In the same way, rd may be normalized to another quantity than the lower threshold Tdm (e.g. the characteristic dark luminance value Ld). The dark visual loudness component Vd is thus obtained as a function of a barycenter of the luminance of the dark pixels with respect to the characteristic dark luminance value Ld and/or of a zero luminance reference value, i.e. in a very simple and straightforward manner from those characteristic values. As for the ratio rnorm, rd may be weighted according to a non-linear function, which may be any function that takes a normalized scalar as input and produces a scalar as output. This includes, but is not limited to power functions, logarithmic functions, exponential functions, trigonometric and sigmoidal functions. The advantage of such weighting is that the contribution to overall visual loudness of the dark component can be better controlled. In that case, the ratio rd weighted according to such non-linear function may be directly interpreted as the dark visual loudness component Vd.
In some embodiments, the multimedia content includes a video content, and in block 100b (figure 1 b), the at least two visual loudness components obtained in block 100a are further adjusted for providing at least two adjusted visual loudness components as a function of the at least two visual loudness components and of at least two previous visual loudness components obtained for a previous image in the video content.
In one embodiment, a leaky integration is applied to the visual loudness components Vd, Vm and Vh and/or to the characteristic luminance values Ld, Lm and Lh obtained for successive images in the video content.
Thus, the disclosed method is temporally stable so that no flicker or other unwanted artefacts are introduced in the video content.
Furthermore, when a sharp transition occurs in the content, or during the time immediately following such a sharp transition (where for instance a transition may be due to switching between a movie and a commercial and back), the method according to this embodiment allows a smooth adaptation of the content so as to follow the speed of human visual adaptation for instance.
For instance, these parameters may be adjusted such that the overall visual loudness V follows a temporal Cornsweet profile (see "Cornsweet, T. (2012). "Visual perception". Academic press"). This could be used to increase the visual appearance of temporal contrast, or alternatively to decrease the visual appearance of temporal contrast.
In block 100c (figure 1 b), the visual loudness value to be obtained from at least the luminance of the current pixel (i.e. the output of block 100) is calculated as a function of the at least two visual loudness components obtained in block 100a.
However, in the embodiments discussed above in relation with block 100b, the multimedia content includes a video content, and the visual loudness value to be obtained is calculated in block 100c as a function of the at least two adjusted visual loudness components obtained in block 100b. In any case, whatever the nature of the multimedia content (video or single image), the visual loudness value may be obtained according to the method disclosed above in relation with block 100 and equation (Eq-1 ) when introducing the representation of visual loudness as pairs of numbers. It means that a curve representation Vjmage(x. ) = Cimage ^image C^ y)) can be derived using for instance an interpolation process based on the at least two pairs of numbers derived from the visual loudness components obtained in block 100a or in block 100c.
The visual loudness value is thus obtained as extrapolated from the at least two visual loudness components (or two adjusted visual loudness components) and from the at least two corresponding characteristic luminance values, based on the luminance Limage(x, y) of the current pixel located at position (x, y).
In block 1 10 (figures 1 a and 1 b), a reference luminance value is associated to the visual loudness value obtained in block 100.
In one embodiment, the reference luminance values are stored in a look-up table with entries corresponding to all possible visual loudness values.
In another embodiment, a reference source of content may be created that has a desirable level of visual loudness. The visual loudness for this reference content is represented by at least two among the three pairs of characteristic luminance values / visual loudness components:
(^d,ref< ^d,ref)
(^m,ref< ^m,ref)
(^h,ref< ^ft.ref)
From these values a curve representation Vrei = cref(Lref) can be derived using the interpolation process disclosed above in relation with block 100 and equation (Eq-1 ). The function cref should be monotonically increasing. It is envisaged that the desired visual loudness for an HDR broadcast channel is encoded with this function, and should be applied to all program material to be broadcast over this channel. It is therefore anticipated that it is possible to construct a reference visual loudness presentation that, if represented by a curve, yields a monotonically increasing curve. This reference curve must be monotonically increasing so that it can be inverted.
Reconsidering the similar representation as constructed for the considered current image in block 100c, namely Vjmage (*. y) = cimage (£image ( y)) (This representation may or may not constitute a monotonically increasing curve), a reference luminance value belonging to the reference source of content is associated to the visual loudness Vjmage (*. y) = cimage (£image (*. y)) according to:
^ref (^image C^' y)) = cref (^mage (^image C^' y) ))
In this embodiment, the reference luminance value being associated to the visual loudness value associated to the current pixel is obtained from a luminance value of a pixel in a reference image.
Thus, the luminance of the considered image in the multimedia content can be adapted so as to present a visual consistency with a reference image in a reference source of content.
In block 130 (figure 1 b), the reference luminance value Lref (Limage(x, y)) is averaged with the luminance value Limage(x, y) of the current pixel in the considered image, providing a modified luminance value.
Indeed, in certain applications it may be desirable to affect a partial match of visual loudness, rather than a full match between source and reference content. This can be achieved by means of averaging the reference luminance value Lref (Limage(x, y)) ar>d the luminance Limage (*. y) °f tne current pixel located at position (x, y) in the considered image, for providing a modified luminance value L'ref (Limage(x, y)) given by:
Figure imgf000021_0001
where and β determine the relative contributions of matched and unmatched luminance values of the considered image. In typical applications, the sum of the weights equal 1 , i.e. α + β = 1.
In block 120 (figures 1 a and 1 b), the luminance of the multimedia content is adapted by attributing the reference luminance value obtained in block 1 10 to the current pixel.
Consequently, the "global" luminance of the source image can be adapted so as to get an adapted image that presents a visual consistency with the selected reference.
In the embodiment where the reference luminance value Lref (Limage(x, y)) is averaged with the luminance value Limage(x, y) of the current pixel in the considered image for providing a modified luminance value in block 130, this is the modified luminance value L'ref (Limage(x, y)) that is attributed to the current pixel. In that case, only a partial adaptation of the luminance of the content is performed so that a smooth adaptation of the source content is obtained.
Referring now to figure 2, we illustrate the structural blocks of an exemplary apparatus that can be used for implementing the method for adapting a luminance of a multimedia content according to any of the embodiments disclosed above in relation with figures 1 a and 1 b. In an embodiment, an apparatus 200 for implementing the disclosed method comprises a non-volatile memory 201 (e.g. a read-only memory (ROM) or a hard disk), a volatile memory 203 (e.g. a random access memory or RAM) and a processor 202. The non-volatile memory 201 is a non-transitory computer-readable carrier medium. It stores executable program code instructions, which are executed by the processor 202 in order to enable implementation of the method described above (method for adapting a luminance of a multimedia content) in its various embodiment disclosed in relationship with figures 1 a and 1 b. Non-volatile memory 201 and volatile memory 203 may be combined into a single entity referred to as memory, configured as combined volatile and non-volatile memory, or just one of the types of memory.
Upon initialization, the aforementioned program code instructions are transferred from the non-volatile memory 201 to the volatile memory 203 so as to be executed by the processor 202. The volatile memory 203 likewise includes registers for storing the variables and parameters required for this execution.
All the steps of the above method for adapting a luminance of a multimedia content may be implemented equally well:
by the execution of a set of program code instructions executed by a reprogrammable computing machine such as a PC type apparatus, a DSP (digital signal processor) or a microcontroller. This program code instructions can be stored in a non- transitory computer-readable carrier medium that is detachable (for example a floppy disk, a CD-ROM or a DVD-ROM) or non-detachable; or
by a dedicated machine or component, such as an FPGA (Field Programmable Gate Array), an ASIC (Application-Specific Integrated Circuit) or any dedicated hardware component.
In other words, the disclosure is not limited to a purely software-based implementation, in the form of computer program instructions, but that it may also be implemented in hardware form or any form combining a hardware portion and a software portion.

Claims

1. A method for adapting a luminance of a multimedia content,
comprising, for a current pixel of an image of said multimedia content:
· obtaining (100) a visual loudness value from at least a luminance value of said current pixel and from at least a luminance value of another pixel of said image;
• associating (1 10) a reference luminance value to said visual loudness value; and
• adapting (120) said luminance of said multimedia content by attributing said reference luminance value to said current pixel;
said visual loudness being a measure representative of a luminance of said image with respect to a spatial configuration of distinct brightness areas in said image.
2. An apparatus for adapting a luminance of a multimedia content, comprising:
- a memory; and
- a processor (202) configured for, for a current pixel of an image of said multimedia content:
• obtaining a visual loudness value from at least a luminance value of said current pixel and from at least a luminance value of another pixel of said image;
• associating a reference luminance value to said visual loudness value; and
• adapting said luminance of said multimedia content by attributing said reference luminance value to said current pixel;
said visual loudness being a measure representative of a luminance of said image with respect to a spatial configuration of distinct brightness areas in said image.
3. A method according to claim 1 or an apparatus according to claim 2,
wherein said obtaining (100) a visual loudness value comprises:
• obtaining (100a) at least two visual loudness components belonging to the group comprising:
o a highlight visual loudness component Vh, representing an assessment of the contribution of highlight pixels in a luminance of said image with respect to a spatial configuration of areas containing said highlight pixels in said image;
o a mid-tone visual loudness component Vm, representing an assessment of the contribution of mid-tone pixels in a luminance of said image with respect to a spatial configuration of areas containing said mid-tone pixels in said image; and o a dark visual loudness component Vd, representing an assessment of the contribution of dark pixels in a luminance of said image with respect to a spatial configuration of areas containing said dark pixels in said image;
• calculating (100c) said visual loudness value as a function of said at least two visual loudness components.
4. A method according to claim 1 or 3, or an apparatus according to claim 2 or 3, wherein said reference luminance value being associated to said visual loudness value is obtained from a luminance value of a pixel in a reference image.
5. A method according to any of claims 1 , 3 or 4 further comprising, or an apparatus according to any of claims 2 to 4 further configured for, averaging (130) said reference luminance value with said luminance value of said current pixel, providing a modified luminance value,
said adapting (120) said luminance of said multimedia content corresponding to attributing said modified luminance value to said current pixel.
6. A method or an apparatus according to claim 3 or 4,
wherein said obtaining (100a) at least two visual loudness components comprises calculating (100a1 ) at least one threshold belonging to the group comprising:
• an upper threshold Tmh; and
• a lower threshold Tdm;
said highlight pixels corresponding to pixels with a luminance above said upper threshold Tmh, said dark pixels corresponding to pixels with a luminance below said lower threshold Tdm and said mid-tone pixels corresponding to pixels with a luminance below said upper threshold Tmh and above said lower threshold Tdm.
7. A method or an apparatus according to claim 6,
wherein said obtaining (100a) at least two visual loudness components comprises deriving (100a2) at least two corresponding characteristic luminance values belonging to the group comprising:
• a characteristic highlight luminance value Lh representative of a luminance of lighter pixels in said image; • a characteristic mid-tone luminance value Lm representative of an average luminance of pixels in said image; and
• a characteristic dark luminance value Ld representative of a luminance of darker pixels in said image;
said upper threshold Tmh being a function of said characteristic highlight luminance value Lh and of said characteristic mid-tone luminance value Lm, and
said lower threshold Tdm being a function of said characteristic mid-tone luminance value Lm and of said characteristic dark luminance value Ld.
8. A method or an apparatus according to claim 7,
wherein said visual loudness value is extrapolated from said at least two visual loudness components and from said at least two corresponding characteristic luminance values.
9. A method or an apparatus according to any one of claims 3 to 8,
wherein said obtaining (100a) at least two visual loudness components comprises:
• determining (100a3) a mask Mh representative of a location of said highlight pixels in said image; and
• splitting (100a4) said mask Mh when a density of said highlight pixels in said mask Mh is below a highlight pixels density threshold,
said splitting providing at least one sub-mask presenting a higher density of highlight pixels than said mask Mh;
said highlight visual loudness component Vh being a function of a density of highlight pixels of said at least one sub-mask and of a size of said at least one sub-mask.
10. A method or an apparatus according to claim 9,
wherein said function is a non-linear function of said size of said at least one sub-mask.
11. Method for adapting a luminance of a multimedia content according to any of the claims 7 to 10,
wherein said mid-tone visual loudness component Vm is a function of a barycenter of a luminance of said mid-tone pixels with respect to said characteristic highlight Lh and/or dark Ld luminance values.
12. A method or an apparatus according to any of claims 7 to 1 1 ,
wherein said dark visual loudness component Vd is a function of a barycenter of a luminance of said dark pixels with respect to said characteristic dark luminance value Ld and/or of a zero luminance reference value.
13. A method or an apparatus according to any of claims 3 to 12,
wherein said multimedia content is a video content,
and wherein said obtaining (100) a visual loudness value further comprises adjusting (100b) visual loudness components for providing at least two adjusted visual loudness components as a function of said at least two visual loudness components and of at least two previous visual loudness components obtained for a previous image in said video content,
said visual loudness value being obtained from said at least two adjusted visual loudness components.
14. A computer program product comprising program code instructions for implementing the method according to at least one of claims 1 or 3 to 13, when said program is executed on a computer or a processor.
15. A non-transitory computer-readable carrier medium storing a computer program product according to claim 14.
PCT/EP2017/077161 2016-10-28 2017-10-24 Method for adapting a luminance of a multimedia content, corresponding computer program product and apparatus WO2018077876A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP16306421.5 2016-10-28
EP16306421.5A EP3316248A1 (en) 2016-10-28 2016-10-28 Method for adapting a luminance of a multimedia content, corresponding computer program product and apparatus

Publications (1)

Publication Number Publication Date
WO2018077876A1 true WO2018077876A1 (en) 2018-05-03

Family

ID=57391922

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2017/077161 WO2018077876A1 (en) 2016-10-28 2017-10-24 Method for adapting a luminance of a multimedia content, corresponding computer program product and apparatus

Country Status (2)

Country Link
EP (1) EP3316248A1 (en)
WO (1) WO2018077876A1 (en)

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040257318A1 (en) * 2001-11-02 2004-12-23 Hiroshi Itoh Image display apparatus
US20090146941A1 (en) * 2005-10-18 2009-06-11 Toshiyuki Fujine Liquid crystal display apparatus
US20150035868A1 (en) * 2013-07-31 2015-02-05 Coretronic Corporation Projection device and luminance control method of frame thereof
US20160254028A1 (en) * 2013-07-30 2016-09-01 Dolby Laboratories Licensing Corporation System and Methods for Generating Scene Stabilized Metadata

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040257318A1 (en) * 2001-11-02 2004-12-23 Hiroshi Itoh Image display apparatus
US20090146941A1 (en) * 2005-10-18 2009-06-11 Toshiyuki Fujine Liquid crystal display apparatus
US20160254028A1 (en) * 2013-07-30 2016-09-01 Dolby Laboratories Licensing Corporation System and Methods for Generating Scene Stabilized Metadata
US20150035868A1 (en) * 2013-07-31 2015-02-05 Coretronic Corporation Projection device and luminance control method of frame thereof

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
ALLRED, S. R.; RADONJIC, A.; GILCHRIST, A. L.; BRAINARD, D. H.: "Lightness perception in high dynamic range images: Local and remote luminance ejects", JOURNAL OF VISION, vol. 12, no. 2, 2012, pages 7
CORNSWEET, T.: "Visual perception", 2012, ACADEMIC PRESS
GILCHRIST, A.; KOSSYFIDIS, C.; BONATO, F.; AGOSTINI, T.; CATALIOTTI, J.; LI, X.; ECONOMOU, E.: "An anchoring theory of lightness perception", PSYCHOLOGICAL REVIEW, vol. 106, no. 4, 1999, pages 795
REINHARD, E.; KHAN, E. A.; AKYUZ, A. O.; JOHNSON, G.: "Color imaging: fundamentals and applications", 2008, CRC PRESS

Also Published As

Publication number Publication date
EP3316248A1 (en) 2018-05-02

Similar Documents

Publication Publication Date Title
Veluchamy et al. Image contrast and color enhancement using adaptive gamma correction and histogram equalization
JP7422833B2 (en) A scalable system for controlling color management including various levels of metadata
CN101505421B (en) Method of high dynamic range compression with detail preservation and noise constraints
US9442562B2 (en) Systems and methods of image processing that adjust for viewer position, screen size and viewing distance
US8090198B2 (en) Image processing apparatus, image display apparatus, and image display method
KR101537510B1 (en) Display management server
JP5611337B2 (en) Zone-based tone mapping
KR101446364B1 (en) Method, apparatus and system for providing color grading for displays
Reinhard et al. Calibrated image appearance reproduction.
KR100612835B1 (en) A method and apparatus for generating user preference data regarding color characteristic of image, and method and apparatus for converting color preference of image using the method and appatatus
US8159616B2 (en) Histogram and chrominance processing
US9087385B2 (en) Method for improving images captured underwater
WO2023098251A1 (en) Image processing method, device, and readable storage medium
Kwon et al. Luminance adaptation transform based on brightness functions for LDR image reproduction
JPWO2013129225A1 (en) Image processing apparatus and method
CN107592517B (en) Skin color processing method and device
CN111080722B (en) Color migration method and system based on significance detection
JP6742417B2 (en) Digital image processing method, associated device, terminal device and computer program
Kwon et al. Radiance map construction based on spatial and intensity correlations between LE and SE images for HDR imaging
WO2018077876A1 (en) Method for adapting a luminance of a multimedia content, corresponding computer program product and apparatus
Veluchamy et al. Fuzzy dissimilarity contextual intensity transformation with gamma correction for color image enhancement
Lee et al. Piecewise tone reproduction for high dynamic range imaging
Cyriac et al. Automatic, viewing-condition dependent contrast grading based on perceptual models
CN116167950B (en) Image processing method, device, electronic equipment and storage medium
Mehmood et al. Perceptual Tone Mapping Model for High Dynamic Range Imaging

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17787444

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17787444

Country of ref document: EP

Kind code of ref document: A1