Shi Mengdi, Zhang Xudong, Dong Yunliu, et al. A light field demosaicing method with double guided filtering[J]. Opto-Electronic Engineering, 2019, 46(12): 180539. doi: 10.12086/oee.2019.180539
Citation: Shi Mengdi, Zhang Xudong, Dong Yunliu, et al. A light field demosaicing method with double guided filtering[J]. Opto-Electronic Engineering, 2019, 46(12): 180539. doi: 10.12086/oee.2019.180539

A light field demosaicing method with double guided filtering

    Fund Project: Supported by National Natural Science Foundation of China (61876057, 61471154)
More Information
  • Aiming at the problem that the light field multi-view image quality is poor which is resulting from the specific lenslet structure of the light field camera and pixel aliasing at the lenslet edge, a light field demosaicing algorithm based on double-guided filtering is proposed. First, the G image is reconstructed by reweighting the gradient based threshold free (GBTF) algorithm with the white image and lenslet mask information. Then, the reconstructed G image is used to double-guide the R/B image for reconstruction. Finally, the reconstructed R, G, and B images are combined into a full color image. The demosaicing result demonstrates that compared with other advanced demosaicing algorithms, the index CPSNR is increased by 1.68%, the index SSIM is increased by 2%, and the light field multi-view image obtained by our method has clear edges and less color artifacts.
  • 加载中
  • [1] Ng R, Levoy M, Brédif M, et al. Light field photography with a hand-held plenoptic camera: CS-TR-2005-02[R]. Stanford: Stanford University, 2005.

    Google Scholar

    [2] 王丽娟, 张骏, 张旭东, 等.局部搜索式的Lytro相机微透镜阵列中心标定[J].光电工程, 2016, 43(11): 19–25. doi: 10.3969/j.issn.1003-501X.2016.11.004

    CrossRef Google Scholar

    Wang L J, Zhang J, Zhang X D, et al. Micro-lens array center calibration via local searching using lytro cameras[J]. Opto-Electronic Engineering, 2016, 43(11): 19–25. doi: 10.3969/j.issn.1003-501X.2016.11.004

    CrossRef Google Scholar

    [3] Chan W S, Lam E Y, Ng M K, et al. Super-resolution reconstruction in a computational compound-eye imaging system[J]. Multidimensional Systems and Signal Processing, 2007, 18(2–3): 83–101. doi: 10.1007/s11045-007-0022-3

    CrossRef Google Scholar

    [4] Bishop T E, Zanetti S, Favaro P. Light field superresolution[C]//Proceedings of 2009 IEEE International Conference on Computational Photography (ICCP), 2009: 1–9.

    Google Scholar

    [5] Georgiev T, Lumsdaine A. Superresolution with plenoptic 2.0 cameras[C]//Signal Recovery and Synthesis 2009, 2009.

    Google Scholar

    [6] Bishop T E, Favaro P. The light field camera: extended depth of field, aliasing, and superresolution[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2012, 34(5): 972–986. doi: 10.1109/TPAMI.2011.168

    CrossRef Google Scholar

    [7] Mitra K, Veeraraghavan A. Light field denoising, light field superresolution and stereo camera based refocussing using a GMM light field patch prior[C]//Proceedings of 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, 2012.

    Google Scholar

    [8] Boominathan V, Mitra K, Veeraraghavan A. Improving resolution and depth-of-field of light field cameras using a hybrid imaging system[C]//Proceedings of 2014 IEEE International Conference on Computational Photography (ICCP), 2014.

    Google Scholar

    [9] Yoon Y, Jeon H G, Yoo D, et al. Learning a deep convolutional network for light-field image super-resolution[C]//Proceedings of 2015 IEEE International Conference on Computer Vision Workshop (ICCVW), 2015.

    Google Scholar

    [10] Wang Y W, Liu Y B, Heidrich W, et al. The light field attachment: turning a DSLR into a light field camera using a low budget camera ring[J]. IEEE Transactions on Visualization and Computer Graphics, 2017, 23(10): 2357–2364. doi: 10.1109/TVCG.2016.2628743

    CrossRef Google Scholar

    [11] 邓武, 张旭东, 熊伟, 等.融合全局与局部视角的光场超分辨率重建[J].计算机应用研究, 2019, 36(5): 1549–1554, 1559.

    Google Scholar

    Deng W, Zhang X D, Xiong W, et al. Light field super-resolution using global and local multi-views[J]. Application Research of Computers, 2019, 36(5): 1549–1554, 1559.

    Google Scholar

    [12] Huang X, Cossairt O. Dictionary learning based color demosaicing for plenoptic cameras[C]//Proceedings of 2014 IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2014: 455–460.

    Google Scholar

    [13] Seifi M, Sabater N, Drazic V, et al. Disparity-guided demosaicking of light field images[C]//Proceedings of 2014 IEEE International Conference on Image Processing, 2015.

    Google Scholar

    [14] Yu Z, Yu J Y, Lumsdaine A, et al. An analysis of color demosaicing in plenoptic cameras[C]//Proceedings of 2012 IEEE Conference on Computer Vision and Pattern Recognition, 2012: 901–908.

    Google Scholar

    [15] Xu S, Zhou Z L, Devaney N. Multi-view image restoration from plenoptic raw images[C]//Computer Vision - ACCV 2014 Workshops, 2014, 9009: 3–15.

    Google Scholar

    [16] Cho H, Yoo H. Masking based demosaicking for image enhancement using plenoptic camera[J]. International Journal of Applied Engineering Research, 2018, 13(1): 273–276.

    Google Scholar

    [17] David P, le Pendu M, Guillemot C. White lenslet image guided demosaicing for plenoptic cameras[C]//Proceedings of the 19th International Workshop on Multimedia Signal Processing (MMSP), 2017.

    Google Scholar

    [18] Adams Jr J E, Hamilton Jr J F. Adaptive color plane interpolation in single sensor color electronic camera: EP0732858A3[P]. 1996-09-18.

    Google Scholar

    [19] Malvar H S, He L W, Cutler R. High-quality linear interpolation for demosaicing of bayer-patterned color images[C]//Proceedings of 2004 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2004.

    Google Scholar

    [20] Zhang L, Wu X L. Color demosaicking via directional linear minimum mean square-error estimation[J]. IEEE Transactions on Image Processing, 2005, 14(12): 2167–2178. doi: 10.1109/TIP.2005.857260

    CrossRef Google Scholar

    [21] Pekkucuksen I, Altunbasak Y. Gradient based threshold free color filter array interpolation[C]//Proceedings of 2010 IEEE International Conference on Image Processing, 2010.

    Google Scholar

    [22] Zhang L, Wu X L, Li X. Color demosaicking by local directional interpolation and nonlocal adaptive thresholding[J]. Journal of Electronic Imaging, 2011, 20(2): 023016. doi: 10.1117/1.3600632

    CrossRef Google Scholar

    [23] Chen X D, He L W, Jeon G, et al. Multidirectional weighted interpolation and refinement method for bayer pattern CFA demosaicking[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2015, 25(8): 1271–1282. doi: 10.1109/TCSVT.2014.2313896

    CrossRef Google Scholar

    [24] Alleysson D, Susstrunk S, Herault J. Linear demosaicing inspired by the human visual system[J]. IEEE Transactions on Image Processing, 2005, 14(4): 439–449. doi: 10.1109/TIP.2004.841200

    CrossRef Google Scholar

    [25] Dubois E. Frequency-domain methods for demosaicking of Bayer-sampled color images[J]. IEEE Signal Processing Letters, 2005, 12(12): 847–850. doi: 10.1109/LSP.2005.859503

    CrossRef Google Scholar

    [26] Leung B, Jeon G, Dubois E. Least-squares luma-chroma demultiplexing algorithm for bayer demosaicking[J]. IEEE Transactions on Image Processing, 2011, 20(7): 1885–1894. doi: 10.1109/TIP.2011.2107524

    CrossRef Google Scholar

    [27] Kiku D, Monno Y, Tanaka M, et al. Residual interpolation for color image demosaicking[C]//Proceedings of 2013 IEEE International Conference on Image Processing, 2013.

    Google Scholar

    [28] Kiku D, Monno Y, Tanaka M, et al. Minimized-laplacian residual interpolation for color image demosaicking[J]. Proceedings of the SPIE, 2014, 9023: 90230L.

    Google Scholar

    [29] Dansereau D G, Pizarro O, Williams S B. Decoding, calibration and rectification for lenselet-based plenoptic cameras[C]//Proceedings of 2013 IEEE Conference on Computer Vision and Pattern Recognition, 2013.

    Google Scholar

    [30] Wanner S, Meister S, Goldluecke B. Datasets and benchmarks for densely sampled 4D light fields[C]//Vision Modeling and Visualization, 2013: 225–226.

    Google Scholar

    [31] Mousnier A, Vural E, Guillemot C. Lytro dataset[EB/OL].[2017-05-03]. http://www.irisa.fr/temics/demos/lightField/index.html.

    Google Scholar

  • Overview: With the development of light field imaging technology, the light field camera, as a new multi-view imaging device, has become popular in computer imaging community. Light field cameras can be divided into three categories: camera array cameras, mask cameras, and microlens cameras. Due to its simple structure and small size, the microlens camera has been widely used. Since the microlens camera uses a single CCD sensor with a color filter array (CFA) to capture the 3D scene information, it can only sample one of the RGB values for each pixel. In order to obtain a high quality light field color image, the light field camera needs to be demosaiced to obtain a full-color image. The demosaicing algorithm of the traditional cameras has been studied for decades, and the corresponding technologies are very mature. Different from the regulars image, every microlens image has aliasing or vignetting effect at the boundary owing to its special structure. Therefore, it is not suitable to directly apply a conventional demosaicing algorithm to the microlens images to obtain a full-color image. In recent years, many light field demosaicing algorithms have been proposed to achieve reasonable results when there is no aliasing or vignetting on the microlens images. However, when there are aliasing and vignetting effects on the microlens images, the performance of these algorithms becomes worse and some terrible phenomena may appear in full-color images, such as image blurring and color artifacts. To solve the above issue, a light field demosaicing algorithm based on double-guided filtering is proposed. Wherein, double-guided filtering refers to using two guiding filters, that is, applying a sparse Laplacian to the input image in the first guiding filtering, and obtaining an output by minimizing sparse Laplacian energy. In the second boot filtering, the output of the first boot filter is used as the input, and a standard Laplacian is applied to the input image by minimizing the standard Laplacian energy, which can effectively preserve the structure of the guided image. First, the G image is reconstructed by reweighting the gradient based threshold free (GBTF) algorithm with the white image and lenslet mask information. Then, the reconstructed G image is used to double-guide the R/B image for reconstruction. Finally, the reconstructed R, G, and B images are combined into a full-color image. The experiments are carried out on the synthetic light field dataset and the real scene light field dataset, respectively, which verify the effectiveness of the proposed algorithm by increasing the index CPSNR by 1.68%, the index SSIM by 2%, comparing with the state of the arts. The light field full-color images obtained by our method have clear edges and less color artifacts.

  • 加载中
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Figures(11)

Tables(2)

Article Metrics

Article views(7424) PDF downloads(3032) Cited by(0)

Access History

Other Articles By Authors

Article Contents

Catalog

    /

    DownLoad:  Full-Size Img  PowerPoint