FUELING THE NEXT ERA OF AR/VR ADVANCEMENTS THROUGH EDGEAI
Keywords:
EdgeAI, Augmented Reality (AR), Virtual Reality (VR)Abstract
The fusion of Edge AI with Augmented Reality/Virtual Reality (AR/VR) presents a dynamic landscape filled with promising opportunities and intricate challenges. This partnership between Edge AI and AR/VR technologies reveals a horizon were achieving real-time performance, minimizing latency, and enhancing user immersion becomes attainable. This article explores the diverse possibilities that emerge when seamlessly integrating Edge AI into AR/VR systems. It emphasizes the potential of processing data locally to enable unprecedented levels of interactivity and responsiveness. Additionally, empowering AR/VR devices to make rapid on-the-edge decisions has the potential to revolutionize user experiences in various domains. The paper delves into various techniques for optimizing Edge AI to run efficiently on the resource-constrained hardware found in AR/VR systems. By addressing these challenges thoughtfully, stakeholders can unlock the transformative potential of this synergy, ensuring that the integration of Edge AI and AR/VR enhances human experiences in a responsible and innovative manner.
References
I. Arun Faisal, T. Waluyo Purboyo, and A. Siswo Raharjo Ansori, “A Review of Accelerometer Sensor and Gyroscope Sensor in IMU Sensors on Motion Capture,” Journal of Engineering and Applied Sciences, vol. 15, no. 3, pp. 826–829, Nov. 2019.
Auria, D.D.; Mauro, D.D.; Calandra, D.M.; Cutugno, F. A 3D audio augmented reality system for cultural heritage management and fruition. J. Digit. Inf. Manag. 2015, 13, 203–209.
Hurtado Soler, A., Botella Nicolás, A. M,. & Martínez Gallego, S. (2022). Virtual and Augmented Reality Applied to the Perception of the Sound and Visual Garden. Educ. Sci., 12, 377. https://doi.org/10.3390/educsci12060377
Aulinas, J., Y. R. Petillot, J. Salvi, and X. Lladó. 2008. “The SLAM Problem: A Survey.” Conference on Artificial Intelligence Research and Development, 11th International Conference of the Catalan Association for Artificial Intelligence, Spain, October 22–24.
Ruihao Li, Sen Wang, and Dongbing Gu. Deepslam: A robust monocular slam system with unsupervised deep learning. IEEE Transactions on Industrial Electronics, 68(4):3577–3587, 2021. doi: 10.1109/TIE. 2020.2982096.
P. L. Mazzeo, D. D’Amico, P. Spagnolo, and C. Distante, ‘‘Deep learning based eye gaze estimation and prediction,’’ in Proc. 6th Int. Conf. Smart Sustain. Technol. (SpliTech), Sep. 2021, pp. 1–6, doi: 10.23919/SpliTech52315.2021.9566413.
Yihua Cheng, Haofei Wang, Yiwei Bao, and Feng Lu. Appearance-based gaze estimation with deep learning: A review and benchmark. arXiv preprint arXiv:2104.12668, 2021.
Sanchez-Riera, J.; Srinivasan, K.; Hua, K.-L.; Cheng, W.-H.; Hossain, M.A.; Alhamid, M.F. Robust rgb-d hand tracking using deep learning priors. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 2289–2301.
Terven, J.; Cordova-Esparza, D. A Comprehensive Review of YOLO: From YOLOv1 and Beyond. arXiv 2023, arXiv:2304.00501.
W. Liu, D. Anguelov, D. Erhan, C. Szegedy, S. Reed, C.-Y. Fu, and A. C. Berg, “SSD: Single shot multibox detector,” in ECCV, 2016.
P. Bharati, A. Pramanik, Deep learning techniques—r-cnn to mask rcnn: A survey, in: Computational Intelligence in Pattern Recognition, Springer, 2020, pp. 657–668.
Olaf Ronneberger, Philipp Fischer, and Thomas Brox. 2015. U-net: Convolutional networks for biomedical image segmentation. In MICCAI. Springer, 234–241.
Gao, C.; Yan, J.; Zhou, S.; Chen, B.; Liu, H. Long short-term memory-based recurrent neural networks for nonlinear target tracking. Signal Process. 2019, 164, 67–73.
B. Veeramani, J. W. Raymond, and P. Chanda, “DeepSort: Deep Convolutional Networks for sorting haploid maize seeds,” BMC Bioinformatics, vol. 19, no. 9, p. 289, 2018.
Lee, B., Jo, Y., Yoo, D. & Lee, J. Recent progresses of near-eye display for AR and VR. In Stella, E. (ed.) Multimodal Sensing and Artificial Intelligence: Technologies and Applications II, vol. 11785, 1178503, DOI: 10.1117/12.2596128.International Society for Optics and Photonics (SPIE, 2021).
Li, Z., Chen, L., Liu, C., Zhang, F., Li, Z., Gao, Y., Ha, Y., Xu, C., Quan, S., & Xu, Y. (2021). Animated 3D human avatars from a single image with GAN-based texture inference. Computers and Graphics, 95, 81–91. https://doi.org/10.1016/j.cag.2021.01.002
Maximo Cobos, Jens Ahrens, Konrad Kowalczyk, and Archontis Politis, “An overview of machine learning and other data-based methods for spatial audio capture, processing, and reproduction,” EURASIP Journal on Audio, Speech, and Music Processing, vol. 2022, no. 1, pp. 1–21, 2022.
Gou, J., Yu, B., Maybank, S.J., Tao, D., 2021. Knowledge distillation: A survey. International Journal of Computer Vision, 1–31.
Hinton, G., Vinyals, O. & Dean, J. (2015). Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531.
Bucilua, C., Caruana, R. & Niculescu-Mizil, A. (2006). Model compression. In: SIGKDD.
S. Han, J. Pool, J. Tran et al., “Learning both Weights and Connections for Efficient Neural Networks,” in Advances in Neural Information Processing Systems 28 (NeurIPS 2015), 2015, pp. 1135–1143.
Dwith Chenna, Edge AI: Quantization as The Key to on-DeviceSmartness, International Journal of Artificial Intelligence & Applications (IJAIAP),1(1), 2023, pp. 58-69.
Downloads
Published
Issue
Section
License
Copyright (c) 2023 Dwith Chenna (Author)

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.