Skip to main content
Log in

Savitzky–Golay Filtering-Based Fusion of Multiple Exposure Images for High Dynamic Range Imaging

  • Original Research
  • Published:
SN Computer Science Aims and scope Submit manuscript

Abstract

The problem of compositing multiple exposure images has attracted lots of researchers, over the past years. It all began with the problem of High Dynamic Range (HDR) imaging, for capturing scenes with vast differences in their dynamic range. Fine details in all the areas in these scenes cannot be captured with one single exposure setting of the camera aperture. This leads to multiple exposure images with each image containing accurate representation of different regions dimly lit, well lit and brightly lit in the scenes. One can make a combined HDR image out of these multiple exposure shots. This combination of multiple exposure shots leads to an image of a higher dynamic range in a different image format which cannot be represented in the traditional Low Dynamic Range (LDR) formats. Moreover HDR images cannot be displayed in traditional display devices suitable for LDR. So these images have to undergo a process called as tone mapping for further converting them to be suitable enough to be represented on usual LDR displays. An approach based on Savitzky–Golay parametric filtering which preserves edges, is proposed which uses filtered multiple exposure images to generate the alpha matte coefficients required for fusing the input multiple exposure set. The coefficients generated in the proposed approach helps in retaining the weak edges and the fine textures which are lost as a result of the under and over exposures. The proposed approach is similar in nature to the bilateral filter-based compositing approach for multiple exposure images in the literature but it is novel, in exploring the possibility of compositing using a parametric filtering approach. The proposed approach performs the fusion in the LDR domain and the fused output can also be displayed using standard LDR image formats on standard LDR displays. A brief comparison of the results generated by the proposed method and various other approaches, including the traditional exposure fusion, tone mapping-based techniques and bilateral filter-based approach is presented where in the proposed method compares well and fares better in majority of the test cases.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Raman S, Chaudhuri S. Bilateral filter based compositing for variable exposure photography. Eurographics 2009.

  2. Gowri BG, Hariharan V, Thara S, Sowmya V, Kumar SS, Soman KP. 2D image data approximation using Savitzky Golay filter—smoothing and differencing. 2013 International MutliConference on Automation, Computing, Communication, Control and Compressed Sensing (iMac4s), Kottayam, 2013;365–371. doi: https://doi.org/10.1109/iMac4s.2013.6526438.

  3. Porter T, Duff T. Compositing digital images. ACM SIGGRAPH Comput Graph. 1984;18:253–9.

    Article  Google Scholar 

  4. Ramakrishnan V, Pete DJ. Non subsampled shearlet transform based fusion of multiple exposure images. SN Comput Sci. 2020;1:326. https://doi.org/10.1007/s42979-020-00343-4.

    Article  Google Scholar 

  5. Debevec MJ. Recoveringhighdynamicrangeradiance maps from photographs. SIGGRAPH; 1997. p. 97.

    Google Scholar 

  6. Bogoni L, Hansen M. Pattern-selective color image fusion. Pattern Recognit. 2001;34:1515–26.

    Article  Google Scholar 

  7. Vonikakis V, Bouzos O, Andreadis I. Multi exposure image fusion based on illumination estimation. Proc SIPA Chania Greece. 2011;22–24:135–42.

    Google Scholar 

  8. Tico M, Gelfand N, Pulli K. Motion-blur-free exposure fusion. In: Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, China, 2010;3321–3324.

  9. Jinno T, Okuda M. Multiple exposure fusion for high dynamic range image acquisition. IEEE Trans Image Process. 2012;21:358–65.

    Article  MathSciNet  Google Scholar 

  10. Shen R, Cheng I, Shi J, Basu A. Generalized random walks for fusion of multi-exposure images. IEEE Trans Image Process. 2011;20:3634–46.

    Article  MathSciNet  Google Scholar 

  11. Li S, Kang X. Fast multi-exposure image fusion with median filter and recursive filter. IEEE Trans Consum Electron. 2012;58:626–32.

    Article  Google Scholar 

  12. Mitianoudis N, Stathaki T. Pixel-based and region-based image fusion schemes using ICA bases. Inf Fusion. 2007;8:131–42.

    Article  Google Scholar 

  13. Brinkmann R. The art and science of digital compositing. San Francisco: Morgan Kaufmann Publishers Inc.; 1999.

    Google Scholar 

  14. Porter T, Duff T. Compositing digital images. SIGGRAPH Comput Graph. 1984;18(3):253–9. https://doi.org/10.1145/964965.808606.

    Article  Google Scholar 

  15. Blinn JF. Compositing. 1. Theory. IEEE Comput Graph Appl. 1994;14(5):83–7. https://doi.org/10.1109/38.310740.

    Article  Google Scholar 

  16. Mann S, Picard RW. On being undigital with digital cameras: extending dynamic range by combining exposed pictures. In: In Proc. of IS & T 48th annual conference 1995;422–428.

  17. Goshtasby A. Fusion of multi-exposure images. Image Vis Comput. 2005;23:611–8.

    Article  Google Scholar 

  18. Raman S, Chaudhuri S. Amatte-less,variationalapproach to automatic scene compositing. ICCV; 2007.

    Google Scholar 

  19. Mertens T, Kautz J, Reeth FV. Exposure fusion. Pacific Graphics; 2007.

    Book  Google Scholar 

  20. Chuang Y, Curless B, Salesin DH, Szeliski R. A bayesian approach to digital matting, vol. 2. Kauai Marriott: CVPR; 2001. p. 264–71.

    Google Scholar 

  21. Guan Y, Chen W, Liang X, Ding Z, Peng Q. Easy matting a stroke based approach for continuous image matting. Eurographics. 2006;25(3):567–76.

    Google Scholar 

  22. Ruzon M, Tomasi C. Alpha estimation in natural images, vol. 1. Hilton Head Island: CVPR; 2000. p. 18–25.

    Google Scholar 

  23. Sun J, Jia J, Tang C, Shum H. Poisson matting. Los Angeles: SIGGRAPH; 2004. p. 315–21.

    Google Scholar 

  24. Wang J, Cohen MF. An iterative optimization approach for unified image segmentation and matting. Beijing: ICCV; 2005. p. 936–43.

    Google Scholar 

  25. Apostoloff N, Fitzgibbon A. Bayesian video matting using learnt image priors, vol. 1. Washington: CVPR; 2004. p. 407–14.

    Google Scholar 

  26. Chuang Y, Agarwala A, Curless B, Salesin DH, Szeliski R. Video matting of complex scenes. San Antonio: SIGGRAPH; 2002. p. 243–8.

    Google Scholar 

  27. McGuire M, Matusik W, Pfister H, Hughes JF, Durand F. Defocus video matting. Los Angeles: SIGGRAPH; 2005. p. 567–76.

    Google Scholar 

  28. Smith A. Alpha and the history of digital compositing. Microsoft Tech Memo; 1995. p. 7.

    Google Scholar 

  29. Michael A. A tone mapping algorithm for high contrast images. In: Proceedings of the 13th Eurographics workshop on Rendering (EGRW ’02). Eurographics Association, Goslar, DEU, 2002;145–156.

  30. Francesco B, Alessandro A, Elena S, Thomas B-R, Patrick L, Marina B, Alan C. Dynamic range compression by differential zone mapping based on psychophysical experiments. In: Proceedings of the ACM Symposium on Applied Perception (SAP ’12). Association for Computing Machinery, New York, NY, USA, 2012;39–46. https://doi.org/10.1145/2338676.2338685

  31. Bruce ND. ExpoBlend: Information preserving exposure blending based on normalized log-domain entropy. Comput Graph. 2014;39:12–23.

    Article  Google Scholar 

  32. Lischinski D, Farbman Z, Uyttendaele M, Szeliski R. Interactive local adjustment of tonal values. ACM Trans Graphics (TOG). 2006;25(3):646–53.

    Article  Google Scholar 

  33. Tan J, Huang Y, Wang K. Logarithmic tone mapping algorithm based on block mapping fusion. In: 2018 International Conference on Audio, Language and Image Processing (ICALIP) 2018; 168–173. IEEE.

  34. Reinhard E, Stark M, Shirley P, Ferwerda J. Photographic tone reproduction for digital images. In: Proceedings of the 29th annual conference on Computer graphics and interactive techniques 2002;pp. 267–276.

  35. Savitzky Golay filter for 2D images http://research.microsoft.com/enus/um/people/jckrumm/SavGol/SavGol.htm, September 2012.

  36. Mantiuk R, Kim K, Rempel A, Heidrich W. HDRVDP-2: a calibrated visual metric for visibility and quality predictions in all luminance conditions ACM Trans. Graph. 2011;30:40. https://doi.org/10.1145/1964921.1964935.

    Article  Google Scholar 

Download references

Acknowledgements

We would like to thank Tom Mertens, Jan Kautz and Frank Van Reeth for providing us the set of multiple exposure house images in Fig. 1. We would like to thank the CAVE Computer Vision Laboratory, Columbia University for making available the multiple exposure images from their database shown in Figures, Figs. 6, 9 and 12.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Vivek Ramakrishnan.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Ethical approval

This article does not contain any studies with human participants or animals performed by any of the authors.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ramakrishnan, V., Pete, D.J. Savitzky–Golay Filtering-Based Fusion of Multiple Exposure Images for High Dynamic Range Imaging. SN COMPUT. SCI. 2, 191 (2021). https://doi.org/10.1007/s42979-021-00594-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s42979-021-00594-9

Keywords

Navigation