Hostname: page-component-89b8bd64d-shngb Total loading time: 0 Render date: 2026-05-08T06:34:25.049Z Has data issue: false hasContentIssue false

Fully-automatic inverse tone mapping algorithm based on dynamic mid-level tone mapping

Published online by Cambridge University Press:  24 February 2020

Gonzalo Luzardo*
Affiliation:
imec-IPI-UGent, Ghent University, Sint-Pietersnieuwstraat 41, 9000Ghent, Belgium Facultad de Ingeniería en Electricidad y Computación, ESPOL Polytechnic University, Campus Gustavo Galindo Km. 30.5 Vía Perimetral, Guayaquil, Ecuador
Jan Aelterman
Affiliation:
imec-IPI-UGent, Ghent University, Sint-Pietersnieuwstraat 41, 9000Ghent, Belgium
Hiep Luong
Affiliation:
imec-IPI-UGent, Ghent University, Sint-Pietersnieuwstraat 41, 9000Ghent, Belgium
Sven Rousseaux
Affiliation:
Vlaamse Radio -en Televisieomroeporganisatie, Auguste Reyerslaan 52, Brussels, Belgium
Daniel Ochoa
Affiliation:
Facultad de Ingeniería en Electricidad y Computación, ESPOL Polytechnic University, Campus Gustavo Galindo Km. 30.5 Vía Perimetral, Guayaquil, Ecuador
Wilfried Philips
Affiliation:
imec-IPI-UGent, Ghent University, Sint-Pietersnieuwstraat 41, 9000Ghent, Belgium
*
Corresponding authors: G. Luzardo Email: GonzaloRaimundo.LuzardoMorocho@UGent.be

Abstract

High Dynamic Range (HDR) displays can show images with higher color contrast levels and peak luminosities than the common Low Dynamic Range (LDR) displays. However, most existing video content is recorded and/or graded in LDR format. To show LDR content on HDR displays, it needs to be up-scaled using a so-called inverse tone mapping algorithm. Several techniques for inverse tone mapping have been proposed in the last years, going from simple approaches based on global and local operators to more advanced algorithms such as neural networks. Some of the drawbacks of existing techniques for inverse tone mapping are the need for human intervention, the high computation time for more advanced algorithms, limited low peak brightness, and the lack of the preservation of the artistic intentions. In this paper, we propose a fully-automatic inverse tone mapping operator based on mid-level mapping capable of real-time video processing. Our proposed algorithm allows expanding LDR images into HDR images with peak brightness over 1000 nits, preserving the artistic intentions inherent to the HDR domain. We assessed our results using the full-reference objective quality metrics HDR-VDP-2.2 and DRIM, and carrying out a subjective pair-wise comparison experiment. We compared our results with those obtained with the most recent methods found in the literature. Experimental results demonstrate that our proposed method outperforms the current state-of-the-art of simple inverse tone mapping methods and its performance is similar to other more complex and time-consuming advanced techniques.

Information

Type
Original Paper
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution, and reproduction in any medium, provided the original work is properly cited.
Copyright
Copyright © The Authors, 2020
Figure 0

Fig. 1. The shape of the proposed inverse tone mapping function. LDR input values are normalized between 0 and 1. The maximum output HDR value depends on the peak luminance in nits of the HDR display e.g. 6000, or 1 for expressing that we intend to achieve the maximum luminance supported by the display.

Figure 1

Fig. 2. Example of LDR images obtained from the HDR-LDR image dataset. As can be seen, the dataset contains images with a wide range of contrast and lighting conditions.

Figure 2

Table 1. Details about the size and number of pairs included in the HDR-LDR Image Dataset.

Figure 3

Table 2. First-order image statistics computed.

Figure 4

Fig. 3. Scheme of the subjective experiment carried out to manually obtain middle-gray out values that best match the luminance channel of the inverse tone-mapped HDR image (EDR) and its corresponding professionally graded HDR image (HDR) in the pair. The middle-gray value ($m_o$) is used by the inverse tone mapping algorithm (iTMO) to compute EDR.

Figure 5

Table 3. Inverse Tone Mapping Algorithms used for the comparison.

Figure 6

Fig. 4. Example of maps generated by HDR-VDP-2.2. The Quality Score (Q-score) is shown on the top of each map.

Figure 7

Fig. 5. Mean Q-scores obtained by HDR-VDP-2.2. Error bars represent the 95% confidence intervals. Horizontal lines represent homogeneous subsets from post-hoc comparisons using the Tukey HSD test.

Figure 8

Fig. 6. Mean Q-scores obtained by HDR-VDP-2.2. Results have been grouped according to which sequence the processed image of the test dataset belongs to. Error bars represent the standard deviation.

Figure 9

Fig. 7. Comparison of different results of inverse tone mapping. An image with a lack of details in saturated areas was used as input. DRTMO and HDRCNN better reconstruct the lost details. Q-score is shown in parentheses. HDR images were tone-mapped by Mantiuk et al. [39] operator.

Figure 10

Fig. 8. Distortion maps generated by DRIM.

Figure 11

Fig. 9. Comparison of different results obtained by each tested algorithm using a highly overexposed image as input. As can be seen, DRTMO and HDRCNN produce artifacts in overexposed areas. For visualization, HDR images were tone-mapped by Mantiuk et al. [39] operator.

Figure 12

Fig. 10. The results of the subjective experiment scaled in JOD units (higher the values, the better). The difference of 1 JOD indicates that 75% of observers selected one condition as better than the other. Absolute values are arbitrary and only the relative differences are relevant. The error bars denote 95% confidence intervals computed by bootstrapping.

Figure 13

Fig. 11. The results of the subjective experiment scaled in JOD units. Points represent conditions and solid lines represent statistically significant differences, as opposed to dashed lines. The x-axis shows the JOD scaling

Figure 14

Table 4. Computation-time for processing one frame ($1920\times 1080$) in milliseconds (ms).