TY - GEN
T1 - Compressive Light Field Reconstructions Using Deep Learning
AU - Gupta, Mayank
AU - Jauhari, Arjun
AU - Kulkarni, Kuldeep
AU - Jayasuriya, Suren
AU - Molnar, Alyosha
AU - Turaga, Pavan
N1 - Funding Information:
al.[31], helpingdesignnewtypesofcodedlightfieldcam-eras. Finally, wecouldexploretherecentunifiednetwork architecture presented by Chang et al. [10] that applies to all inverseproblemsoftheformy = Ax. Whileourworkhas focusedonprocessingsingleframesoflightfieldvideoef-ficiently,wecouldexploreperformingcodingjointlyinthe spatio-angulardomainandtemporaldomain. Thiswould helpimprovethe compressionratioforthese sensors, and potentiallyleadtolightfieldvideothatiscapturedatinter-active (1-15 FPS) frame rates. Finally, it would be interest-ingtoperforminferenceoncompressedlightfieldmeasure-mentsdirectly(similartotheworkforinferenceon2Dcom-pressedimages[29,22])thataimstoextractmeaningfulse-manticinformation. Allofthesefuturedirectionspointto Acknowledgements: Theauthorswouldliketothank theanonymousreviewers fortheirdetailedfeedback, Siva Sankalpforrunningsomeexperiments, andMarkBuckler forGPUcomputingsupport. AJwassupportedbyagift fromQualcomm. KKandPTwerepartiallysupportedby NSF CAREER grant 1451263. SJ was supported by a NSF Graduate Research Fellowship and a Qualcomm Innovation Fellowship.
Publisher Copyright:
© 2017 IEEE.
PY - 2017/8/22
Y1 - 2017/8/22
N2 - Light field imaging is limited in its computational processing demands of high sampling for both spatial and angular dimensions. Single-shot light field cameras sacrifice spatial resolution to sample angular viewpoints, typically by multiplexing incoming rays onto a 2D sensor array. While this resolution can be recovered using compressive sensing, these iterative solutions are slow in processing a light field. We present a deep learning approach using a new, two branch network architecture, consisting jointly of an autoencoder and a 4D CNN, to recover a high resolution 4D light field from a single coded 2D image. This network decreases reconstruction time significantly while achieving average PSNR values of 26-32 dB on a variety of light fields. In particular, reconstruction time is decreased from 35 minutes to 6.7 minutes as compared to the dictionary method for equivalent visual quality. These reconstructions are performed at small sampling/compression ratios as low as 8%, allowing for cheaper coded light field cameras. We test our network reconstructions on synthetic light fields, simulated coded measurements of real light fields captured from a Lytro Illum camera, and real coded images from a custom CMOS diffractive light field camera. The combination of compressive light field capture with deep learning allows the potential for real-time light field video acquisition systems in the future.
AB - Light field imaging is limited in its computational processing demands of high sampling for both spatial and angular dimensions. Single-shot light field cameras sacrifice spatial resolution to sample angular viewpoints, typically by multiplexing incoming rays onto a 2D sensor array. While this resolution can be recovered using compressive sensing, these iterative solutions are slow in processing a light field. We present a deep learning approach using a new, two branch network architecture, consisting jointly of an autoencoder and a 4D CNN, to recover a high resolution 4D light field from a single coded 2D image. This network decreases reconstruction time significantly while achieving average PSNR values of 26-32 dB on a variety of light fields. In particular, reconstruction time is decreased from 35 minutes to 6.7 minutes as compared to the dictionary method for equivalent visual quality. These reconstructions are performed at small sampling/compression ratios as low as 8%, allowing for cheaper coded light field cameras. We test our network reconstructions on synthetic light fields, simulated coded measurements of real light fields captured from a Lytro Illum camera, and real coded images from a custom CMOS diffractive light field camera. The combination of compressive light field capture with deep learning allows the potential for real-time light field video acquisition systems in the future.
UR - http://www.scopus.com/inward/record.url?scp=85030242715&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85030242715&partnerID=8YFLogxK
U2 - 10.1109/CVPRW.2017.168
DO - 10.1109/CVPRW.2017.168
M3 - Conference contribution
AN - SCOPUS:85030242715
T3 - IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
SP - 1277
EP - 1286
BT - Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2017
PB - IEEE Computer Society
T2 - 30th IEEE Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2017
Y2 - 21 July 2017 through 26 July 2017
ER -