Image fusion via non-local sparse K-SVD dictionary learning

Ying Li, Fangyi Li, Bendu Bai, Qiang Shen

Research output: Contribution to journalArticlepeer-review

30 Citations (SciVal)
265 Downloads (Pure)

Abstract

Image fusion aims to merge two or more images captured via various sensors of the same scene to construct a more informative image by integrating their details. Generally, such integration is achieved through the manipulation of the representations of the images concerned. Sparse representation plays an important role in the effective description of images, offering a great potential in a variety of image processing tasks, including image fusion. Supported by sparse representation, in this paper, an approach for image fusion by the use of a novel dictionary learning scheme is proposed. The nonlocal self-similarity property of the images is exploited, not only at the stage of learning the underlying description dictionary but during the process of image fusion. In particular, the property of nonlocal self-similarity is combined with the traditional sparse dictionary. This results in an improved learned dictionary, hereafter referred to as the nonlocal sparse K-SVD dictionary (where K-SVD stands for the K times singular value decomposition that is commonly used in the literature), and abbreviated to NL_SK_SVD. The performance of the NL_SK_SVD dictionary is applied for image fusion using simultaneous orthogonal matching pursuit. The proposed approach is evaluated with different types of images, and compared with a number of alternative image fusion techniques. The resultant superior fused images using the present approach demonstrates the efficacy of the NL_SK_SVD dictionary in sparse image representation.
Original languageEnglish
Pages (from-to)1814-1823
JournalApplied Optics
Volume55
Issue number7
Early online date22 Jan 2016
DOIs
Publication statusPublished - 01 Mar 2016

Fingerprint

Dive into the research topics of 'Image fusion via non-local sparse K-SVD dictionary learning'. Together they form a unique fingerprint.

Cite this