Abstract
This study proposes the method of measuring 3D object shapes in an immersive space using a motion capture system. We report on the visualizing the distortion of acrylic panels mounted on a large aerial display and measuring the aberration of the aerial image using a motion capture system. Large aerial displays are made of large acrylic panels, which are subject to distortion due to their own weight. We succeeded in visualizing the shape of the acrylic plate by motion capture and 3D plotting of the positional information. Using a motion capture system, it was found that the aerial image formed by the distorted acrylic plate exhibits astigmatism, which is the difference between the vertical and horizontal focusing position. Furthermore, by drawing the shape of the side surface of the acrylic plate using poster papers, the coordinates were extracted from the imitation paper image, the radius of curvature of the acrylic plate was calculated, and the aberration was calculated. It was found that it is possible to measure the shape in an immersive space using the motion capture.
Similar content being viewed by others
Data availability
The datasets generated and/or analyzed during the current study are available from the corresponding author on reasonable request.
References
Pereira, M.F., Prahm, C., Kolbenschlag, J., Oliveira, E., Rodrigues, N.F.: Application of AR and VR in hand rehabilitation: a systematic review. J. Biomed. Inform. 111, 103584 (2020)
Yamamoto, H., Tomiyama, Y., Suyama, S.: Floating aerial LED signage based on aerial imaging by retro-reflection (AIRR). Opt. Express 22, 26919–26924 (2014)
Terashima, Y., Suyama, S., Yamamoto, H.: Aerial depth-fused 3D image formed with aerial imaging by retro-reflection (AIRR). Opt. Rev. 26, 179–186 (2019)
Omoto, T., Fujii, K., Yasugi, M., Suyama, S., Yamamoto, H.: 3D Aerial display combining optical see-through aerial imaging by retro-reflection with depth-fused 3D display. Proc. IDW ‘22, pp. 379–382 (2022)
Abe, E., Yasugi, M., Takeuchi, H., Watanabe, E., Kamei, Y., Yamamoto, H.: Development of omnidirectional aerial display with aerial imaging by retro-reflection (AIRR) for behavioral biology experiments. Opt. Rev. 26, 221–229 (2019)
Yasui, M., Alvissalim, M., Yamamoto, H., Ishikawa, M.: Immersive 3D environment by floating display and high-speed gesture UI integration. SICE 52, 134–140 (2016)
Kikuta, H., Yasugi, M., Yamamoto, H.: Examination of deblur processing according optical parameters in aerial image. Optcon 1(3), 462–474 (2022)
Fujii, K., Yasugi, M., Maekawa, S., Yamamoto, H.: Aerial imaging steganography method for aerial imaging by retro-reflection with dual acrylic ball. Opt. Rev. 26, 250–260 (2022)
Yasugi, M., Yamamoto. H.: Exploring the combination of optical components suitable for the large device to form aerial image by AIRR,” Proc. IDW ‘19, pp. 1382–1383 (2019)
Kakinuma, R., Yasugi, M., Ito, S., Fujii, K., Yamamoto, H.: Aerial interpersonal 3D screen with AIRR that shares your gesture and your screen with an opposite viewer. Proc. IMID 2018, 636 (2018)
Yasugi, M., Yamamoto, H.: Optical design suitable for both immersive aerial display system and capturing user motion. Proc. IDW ‘21, pp. 733–735 (2021)
Yasugi, M., Adachi, M., Inoue, K., Ninomiya, N., Suyama, S., Yamamoto, H.: Development of aerial interface by integrating omnidirectional aerial display, motion tracking, and virtual reality space construction. J. Robot. Mechatron. 34(5), 1175–1183 (2022)
Adachi, M., Inoue, K., Yasugi, M., Ninomiya, N., Suyama, S., Yamamoto, H.: Evaluation of response time of AIRR with immersive aerial interface by 3D motion capture. Proc. IDW ‘22, pp. 792–795 (2022)
Inoue, K., Yasugi, M., Yamamoto, H., Ninomiya, N.: Improvement of the distortion of aerial displays and proposal for utilizing distortion to emulate three-dimensional aerial image. Opt. Rev. 29, 261–266 (2022)
Adachi, M., Yasugi, M., Suyama, S., Yamamoto, H.: Astigmatism on 3D image position in a large aerial display measured by motion capture. The 12th Laser Display and Lightning Conference 2023 (LDC 2023), p. LDC9–03 (2023)
Funding
A part of this work was supported by JSPS KAKENHI (19H04155, 20H05702).
Author information
Authors and Affiliations
Contributions
M.A. contributed for this paper as 1st author. She developed the experimental system and conducted the experiments and wrote the original draft. M.Y. and S.S. designed the experimental system and edited the manuscript. H.Y. designed the experiments and edited the manuscript.
Corresponding author
Ethics declarations
Conflicts of interest
The authors declare no conflicts of interest associated with this manuscript.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Adachi, M., Yasugi, M., Suyama, S. et al. Method of acquiring shapes using motion capture of aerial images formed by large acrylic panels. Opt Rev 30, 647–656 (2023). https://doi.org/10.1007/s10043-023-00844-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10043-023-00844-6