Integrated Multi-view 3D Image Capture and Motion Parallax 3D Display System
We propose an integrated 3D image capture and display system using a transversely moving camera, regular 2D display screen and user tracking that can facilitate the multi-view capture of a real scene or object and display the captured perspective views in 3D. The motion parallax 3D technique is used to capture the depth information of the object and display the corresponding views to the user using head tracking. The system is composed of two parts, the first part consists of a horizontally moving camera interfaced with a customized camera control and capture application. The second part consist of a regular LCD screen combined with web camera and user tracking application. The 3D multi-view images captured through the imaging setup are relayed to the display based on the user location and corresponding view is dynamically displayed on the screen based on the viewing angle of the user with respect to the screen. The developed prototype system provides the multi-view capture of 60 views with the step size of 1 cm and greater than 40˚ field-of-view overlap. The display system relays 60 views providing the viewing angle coverage of ±35˚ where the angular difference between two views is 1.2˚.
B. Rogers and M. Graham, Similarities between motion parallax and stereopsis in human depth perception, Vision Res., vol. 22, no. 2, pp. 261–270, 1982. DOI: https://doi.org/10.1016/0042-6989(82)90126-2
J. Geng, Three-dimensional display technologies, Adv. Opt. Photonics, vol. 5, no. 4, pp. 456–535, 2013. DOI: https://doi.org/10.1364/AOP.5.000456
B. Rogers and M. Graham, Motion parallax as an independent cue for depth perception, Perception, vol. 8, no. 2, pp. 125–134, 1979. DOI: https://doi.org/10.1068/p080125
R. Kumar, Smartphone 3D Camera Market by Technology, 2018. [Online]. Available: https://www.alliedmarketresearch.com/smartphone-3d-camera-market. [Accessed: 02-Feb-2021].
W. Matusik and H. Pfister, 3D TV: A Scalable System for Real-Time Acquisition, Transmission, and Autostereoscopic Display of Dynamic Scenes, ACM SIGGRAPH 2004 Pap. - SIGGRAPH ’04, vol. 23, no. 3, p. 814, 2004. DOI: https://doi.org/10.1145/1015706.1015805
B. Wilburn et al., High performance imaging using large camera arrays, ACM Trans. Graph., vol. 24, no. 3, p. 765, 2005. DOI: https://doi.org/10.1145/1073204.1073259
H. Navarro, J. C. Barreiro, G. Saavedra, and B. Javidi, High-resolution far-field integral-imaging camera by double snapshot, Opt. Express, vol. 20, no. 2, pp. 890–895, 2012. DOI: https://doi.org/10.1364/OE.20.000890
V. J. Traver, P. Latorre-carmona, E. Salvador-balaguer, F. Pla, and B. Javidi, Three-Dimensional Integral Imaging for Gesture Recognition Under Occlusions, vol. 24, no. 2, pp. 171–175, 2017. DOI: https://doi.org/10.1109/LSP.2016.2643691
J. Kim, J.-H. Jung, C. Jang, and B. Lee, Real-time capturing and 3D visualization method based on integral imaging, Opt. Express, vol. 21, no. 16, pp. 18742–18753, 2013. DOI: https://doi.org/10.1364/OE.21.018742
Z. Cai, X. Liu, G. Pedrini, W. Osten, and X. Peng, Structured-light-field 3D imaging without phase unwrapping, vol. 129, no. January, 2020. DOI: https://doi.org/10.1016/j.optlaseng.2020.106047
V. Ganapathi, C. Plagemann, D. Koller, and S. Thrun, Real time motion capture using a single time-of-flight camera, in IEEE Conference on Computer Vision and Pattern Recognition, 2010, pp. 755–762. DOI: https://doi.org/10.1109/CVPR.2010.5540141
R. Schulein, M. Daneshpanah, and B. Javidi, 3D imaging with axially distributed sensing, Opt. Lett., vol. 34, no. 13, pp. 2012–2014, 2012. DOI: https://doi.org/10.1364/OL.34.002012
Y. Piao, H. Qu, M. Zhang, and M. Cho, Three-dimensional integral imaging display system via off-axially distributed image sensing, Opt. Lasers Eng., vol. 85, pp. 18–23, 2016. DOI: https://doi.org/10.1016/j.optlaseng.2016.03.023
D. Dunn et al., Wide Field of View Varifocal Near-Eye Display Using See-Through Deformable Membrane Mirrors, IEEE Trans. Vis. Comput. Graph., vol. 23, no. 4, pp. 1275–1284, 2017. DOI: https://doi.org/10.1109/TVCG.2017.2657058
M. Date, S. Shimizu, D. Mikami, and Y. Kusachi, Smooth motion parallax and high resolution display based on visually equivalent light field 3D, vol. 1140205, no. April 2020, p. 4, 2020. DOI: https://doi.org/10.1117/12.2557356
H. Watanabe, N. Okaichi, T. Omura, M. Kano, H. Sasaki, and M. Kawakita, Aktina Vision: Full-parallax three-dimensional display with 100 million light rays, Sci. Rep., vol. 9, no. 1, pp. 1–9, 2019. DOI: https://doi.org/10.1038/s41598-019-54243-6
P. Wang et al., Demonstration of a low-crosstalk super multi-view light field display with natural depth cues and smooth motion parallax, Opt. Express, vol. 27, no. 23, p. 34442, 2019. DOI: https://doi.org/10.1364/OE.27.034442
A. Maimone, A. Georgiou, and J. S. Kollin, Holographic near-eye displays for virtual and augmented reality, ACM Trans. Graph., vol. 36, no. 4, pp. 1–16, 2017. DOI: https://doi.org/10.1145/3072959.3073624
K.-H. L. Son, Jung-Young, Hyoung Lee, Beom-Ryeol Lee, Holographic and Light-Field Imaging as Future 3-D Displays, Proc. IEEE, vol. 105, no. 5, pp. 789–804, 2017. DOI: https://doi.org/10.1109/JPROC.2017.2666538
M. H. Abdurrahman, H. Amran Darwito, and A. Saleh, Face Recognition System for Prevention of Car Theft with Haar Cascade and Local Binary Pattern Histogram using Raspberry Pi, Emit. Int. J. Eng. Technol., vol. 8, no. 2, pp. 407–425, 2020. DOI: https://doi.org/10.24003/emitter.v8i2.534
C. Li, Z. Qi, N. Jia, and J. Wu, Human face detection algorithm via Haar cascade classifier combined with three additional classifiers, ICEMI 2017 - Proc. IEEE 13th Int. Conf. Electron. Meas. Instruments, vol. 2018-Janua, pp. 483–487, 2017.
Copyright (c) 2022 EMITTER International Journal of Engineering Technology
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License.
The copyright to this article is transferred to Politeknik Elektronika Negeri Surabaya(PENS) if and when the article is accepted for publication. The undersigned hereby transfers any and all rights in and to the paper including without limitation all copyrights to PENS. The undersigned hereby represents and warrants that the paper is original and that he/she is the author of the paper, except for material that is clearly identified as to its original source, with permission notices from the copyright owners where required. The undersigned represents that he/she has the power and authority to make and execute this assignment. The copyright transfer form can be downloaded here .
The corresponding author signs for and accepts responsibility for releasing this material on behalf of any and all co-authors. This agreement is to be signed by at least one of the authors who have obtained the assent of the co-author(s) where applicable. After submission of this agreement signed by the corresponding author, changes of authorship or in the order of the authors listed will not be accepted.
Retained Rights/Terms and Conditions
- Authors retain all proprietary rights in any process, procedure, or article of manufacture described in the Work.
- Authors may reproduce or authorize others to reproduce the work or derivative works for the author’s personal use or company use, provided that the source and the copyright notice of Politeknik Elektronika Negeri Surabaya (PENS) publisher are indicated.
- Authors are allowed to use and reuse their articles under the same CC-BY-NC-SA license as third parties.
- Third-parties are allowed to share and adapt the publication work for all non-commercial purposes and if they remix, transform, or build upon the material, they must distribute under the same license as the original.
Plagiarism screening will be conducted by EMITTER Journal Editorial Board using iThenticate Plagiarism Checker and CrossCheck plagiarism screening service. The author should download and sign the declaration of plagiarism form here and resubmit it with the copyright transfer form via online submission.