Journal Home > Volume 2 , Issue 1

The ability of light gathering of plenoptic camera opens up new opportunities for a wide range of computer vision applications. An efficient and accurate method to calibrate plenoptic camera is crucial for its development. This paper describes a 10-intrinsic-parameter model for focused plenoptic camera with misalignment. By exploiting the relationship between the raw image features and the depth-scale information in the scene, we propose to estimate the intrinsic parameters from raw images directly, with a parallel biplanar board which provides depth prior. The proposed method enables an accurate decoding of light field on both angular and positional information, and guarantees a unique solution for the 10 intrinsic parameters in geometry. Experiments on both simulation and real scene data validate the performance of the proposed calibration method.


menu
Abstract
Full text
Outline
About this article

Decoding and calibration method on focused plenoptic camera

Show Author's information Chunping Zhang1( )Zhe Ji1Qing Wang1
School of Computer Science, Northwestern Polytechnical University, Xi’an 710072, China.

Abstract

The ability of light gathering of plenoptic camera opens up new opportunities for a wide range of computer vision applications. An efficient and accurate method to calibrate plenoptic camera is crucial for its development. This paper describes a 10-intrinsic-parameter model for focused plenoptic camera with misalignment. By exploiting the relationship between the raw image features and the depth-scale information in the scene, we propose to estimate the intrinsic parameters from raw images directly, with a parallel biplanar board which provides depth prior. The proposed method enables an accurate decoding of light field on both angular and positional information, and guarantees a unique solution for the 10 intrinsic parameters in geometry. Experiments on both simulation and real scene data validate the performance of the proposed calibration method.

Keywords: calibration, focused plenoptic camera, depth prior, intrinsic parameters

References(20)

[1]
Ng, R. Digital light field photography. Ph.D. Thesis. Stanford University, 2006.
[2]
Ng, R.; Levoy, M.; Bredif, M.; Duval, G.; Horowitz, M.; Hanrahan, P. Light field photography with a hand-held plenoptic camera. Stanford University Computer Science Tech Report CSTR 2005-02, 2005.
[3]
Georgiev, T. G.; Lumsdaine, A. Focused plenoptic camera and rendering. Journal of Electronic Imaging Vol. 19, No. 2, 021106, 2010.
[4]
Lumsdaine, A.; Georgiev, T. Full resolution lightfield rendering. Technical Report. Indiana University and Adobe Systems, 2008.
[5]
Lumsdaine, A.; Georgiev, T. The focused plenoptic camera. In: Proceedings of IEEE International Conference on Computational Photography, 1-8, 2009.
DOI
[6]
Dansereau, D. G.; Pizarro, O.; Williams, S. B. Decoding, calibration and rectification for lenselet-based plenoptic cameras. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, 1027-1034, 2013.
DOI
[7]
Perwaß, C.; Wietzke, L. Single lens 3D-camera with extended depth-of-field. In: Proceedings of SPIE 8291, Human Vision and Electronic Imaging XVII, 829108, 2012.
DOI
[8]
Bishop, T. E.; Favaro, P. Plenoptic depth estimation from multiple aliased views. In: Proceedings of IEEE 12th International Conference on Computer Vision Workshops, 1622-1629, 2009.
DOI
[9]
Levoy, M.; Hanrahan, P. Light field rendering. In: Proceedings of the 23rd Annual Conference on Computer Graphics and Interactive Techniques, 31-42, 1996.
DOI
[10]
Wanner, S.; Fehr, J.; Jähne, B. Generating EPI representations of 4D light fields with a single lens focused plenoptic camera. In: Lecture Notes in Computer Science, Vol. 6938. Bebis, G.; Boyle, R.; Parvin, B. et al. Eds. Springer Berlin Heidelberg, 90-101, 2011.
[11]
Hahne, C.; Aggoun, A.; Haxha, S.; Velisavljevic, V.; Fernández, J. C. J. Light field geometry of a standard plenoptic camera. Optics Express Vol. 22, No. 22, 26659-26673, 2014.
[12]
Johannsen, O.; Heinze, C.; Goldluecke, B.; Perwaß, C. On the calibration of focused plenoptic cameras. In: Lecture Notes in Computer Science, Vol. 8200. Grzegorzek, M.; Theobalt, C.; Koch, R.; Kolb, A. Eds. Springer Berlin Heidelberg, 302-317, 2013.
DOI
[13]
Georgiev, T.; Lumsdaine, A.; Goma, S. Plenoptic principal planes. Imaging Systems and Applications, OSA Technical Digest (CD), paper JTuD3, 2011.
[14]
Hahne, C.; Aggoun, A.; Velisavljevic, V. The refocusing distance of a standard plenoptic photograph. In: Proceedings of 3DTV-Conference: The True Vision—Capture, Transmission and Display of 3D Video, 1-4, 2015.
DOI
[15]
Birklbauer, C.; Bimber, O. Panorama light-field imaging. Computer Graphics Forum Vol. 33, No. 2, 43-52, 2014.
[16]
Bok, Y.; Jeon, H.-G.; Kweon, I. S. Geometric calibration of micro-lens-based light-field cameras using line features. In: Lecture Notes in Computer Science, Vol. 8694. Fleet, D.; Pajdla, T.; Schiele, B.; Tuytelaars, T. Eds. Springer International Publishing, 47-61, 2014.
[17]
Thomason, C. M.; Thurow, B. S.; Fahringer, T. W. Calibration of a microlens array for a plenoptic camera. In: Proceedings of the 52nd Aerospace Sciences Meeting, AIAA SciTech, AIAA 2014-0396, 2014.
DOI
[18]
Zhang, Z. A flexible new technique for camera calibration. IEEE Transactions on Pattern Analysis and Machine Intelligence Vol. 22, No. 11, 1330-1334, 2000.
[19]
Cho, D.; Lee, M.; Kim, S.; Tai, Y.-W. Modeling the calibration pipeline of the Lytro camera for high quality light-field image reconstruction. In: Proceedings of IEEE International Conference on Computer Vision, 3280-3287, 2013.
DOI
[20]
Sabater, N.; Drazic, V.; Seifi, M.; Sandri, G.; Perez, P. Light-field demultiplexing and disparity estimation. 2014. Available at https://hal.archives-ouvertes.fr/hal-00925652/document.
Publication history
Copyright
Acknowledgements
Rights and permissions

Publication history

Revised: 01 December 2015
Accepted: 13 January 2016
Published: 12 March 2016
Issue date: March 2016

Copyright

© The Author(s) 2016

Acknowledgements

The work is supported by the National Natural Science Foundation of China (Nos. 61272287 and 61531014) and the research grant of State Key Laboratory of Virtual Reality Technology and Systems (No. BUAAVR-15KF-10).

Rights and permissions

This article is published with open access at Springerlink.com

The articles published in this journal are distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Other papers from this open access journal are available free of charge from http://www.springer.com/journal/41095. To submit a manuscript, please go to https://www.editorialmanager.com/cvmj.

Return