
Citation: Chuanlong Li, Xingming Sun, Yuqian Li. Information hiding based on Augmented Reality[J]. Mathematical Biosciences and Engineering, 2019, 16(5): 4777-4787. doi: 10.3934/mbe.2019240
[1] | Xianyi Chen, Anqi Qiu, Xingming Sun, Shuai Wang, Guo Wei . A high-capacity coverless image steganography method based on double-level index and block matching. Mathematical Biosciences and Engineering, 2019, 16(5): 4708-4722. doi: 10.3934/mbe.2019236 |
[2] | Dingwei Tan, Yuliang Lu, Xuehu Yan, Lintao Liu, Longlong Li . High capacity reversible data hiding in MP3 based on Huffman table transformation. Mathematical Biosciences and Engineering, 2019, 16(4): 3183-3194. doi: 10.3934/mbe.2019158 |
[3] | Yongju Tong, YuLing Liu, Jie Wang, Guojiang Xin . Text steganography on RNN-Generated lyrics. Mathematical Biosciences and Engineering, 2019, 16(5): 5451-5463. doi: 10.3934/mbe.2019271 |
[4] | Jimmy Ming-Tai Wu, Jerry Chun-Wei Lin, Philippe Fournier-Viger, Youcef Djenouri, Chun-Hao Chen, Zhongcui Li . The density-based clustering method for privacy-preserving data mining. Mathematical Biosciences and Engineering, 2019, 16(3): 1718-1728. doi: 10.3934/mbe.2019082 |
[5] | Wu Zeng . Image data augmentation techniques based on deep learning: A survey. Mathematical Biosciences and Engineering, 2024, 21(6): 6190-6224. doi: 10.3934/mbe.2024272 |
[6] | Luqi Li, Yunkai Zhai, Jinghong Gao, Linlin Wang, Li Hou, Jie Zhao . Stacking-BERT model for Chinese medical procedure entity normalization. Mathematical Biosciences and Engineering, 2023, 20(1): 1018-1036. doi: 10.3934/mbe.2023047 |
[7] | Xinran Zhang, Yongde Zhang, Jianzhi Yang, Haiyan Du . A prostate seed implantation robot system based on human-computer interactions: Augmented reality and voice control. Mathematical Biosciences and Engineering, 2024, 21(5): 5947-5971. doi: 10.3934/mbe.2024262 |
[8] | Cheonshik Kim, Dongkyoo Shin, Ching-Nung Yang . High capacity data hiding with absolute moment block truncation coding image based on interpolation. Mathematical Biosciences and Engineering, 2020, 17(1): 160-178. doi: 10.3934/mbe.2020009 |
[9] | Shuai Cao, Biao Song . Visual attentional-driven deep learning method for flower recognition. Mathematical Biosciences and Engineering, 2021, 18(3): 1981-1991. doi: 10.3934/mbe.2021103 |
[10] | Xiaoyan Zhao, Sanqing Ding . A multi-dimension information fusion-based intelligent prediction approach for health literacy. Mathematical Biosciences and Engineering, 2023, 20(10): 18104-18122. doi: 10.3934/mbe.2023804 |
Augmented reality (AR) has drawn lots of attention in both research field and industry field. AR is a technique which integrates the live view of a camera with virtual digital objects, and displays the object in the real-world environment through the camera [1,2]. This paper presents a novel concept which combines augmented reality with information hiding. Steganography and watermark are two common ways to realize secret communication [3]. Both of these two methods conceal the secret information inside another carrier. The carrier is usually an image with high texture complexity which can minimize the distortion on the original image. Traditional steganographic algorithms typically select secret embedding position using human metrics and often suffer low embedding rate and high detection rate. More recently, there are researchers propose the method of using deep learning to realize image steganography. Deep learning has proven to be very effective among computer vision field. The emergence of the convolutional neural network gives machine the power to extract more useful image features [4]. Li et al. introduce the method of using generative adversarial network (GAN)-generated texture image as the cover image and hide the secret information into it [3]. By deploying the information hiding neural network on the mobile device and integrating the neural network with augmented reality, the secret information can be just hidden in the real world. The approach proposed in this paper utilizes the camera of the augmented reality device for detecting objects in the real world, and then hides the secret message according to the detected object. The communication parties can only conceal or reveal the secret using the system installed on the augmented reality device. By integrating augmented reality, the secret embedding rate can be improved dramatically in potential. The highlights of this paper are summarized as follows: The application of Augmented Reality in academic fields; the combination of Augmented Reality, information hiding and deep learning techniques; the extension of traditional information hiding matching policy; the practical application of coverless information hiding; a potential approach to improve the secret transmission rate.
This paper focuses on the application of augmented reality in conjunction with information hiding instead of the internal augmented reality concept. Section 2 summarizes some related researches about information hiding, deep learning, and augmented reality and puts forward the proposed augmented reality based information hiding architecture in Section 3. Section 4 shows the experimental results and Section 5 concludes this paper.
The applications of augmented reality mostly focus on entertainment currently, the combination of augmented reality with academia is still in its infancy. Here we mainly consider some of the works related with steganography using deep learning and some convolutional neural network best practice since this is the principal part of the whole architecture.
Convolutional neural network is a revolutionary technology in machine learning field, it typically consists of one input layer and one output layer and includes multiple hidden layers between the input and output layer [5]. The convolutional layer applies a convolution kernel over the input tensor. The convolutional kernel is essentially a cross-correlation mathematical operation which can reduce huge amount of network parameters while achieve the same goal as the traditional fully connect network. Krizhevsky et al. first introduced deep convolutional neural network to classify the ImageNet dataset and outclassed the traditional classification methods [5]. A large number of novel applications using CNN springed up after their successful work. CNN is popular used in image classification, image object detection, image object segmentation, image style transfer and so on in computer vision field [6,7,8,9].
Rahim et al. introduce the method of using convolutional neural network based encoder-decoder networks for image steganography [10]. The encoder takes two images as input, one serves as cover image and the other serves as the secret message. The encoder encodes the secret message into the cover image and generates a stego image which resembles the cover image. The decoder accepts the stego image as input and restores the secret. Baluja also proposes a similar auto-encoder architecture as the one of Rehman et al., he points out that the secret image is scattered among all the image channel [11]. Li et al. introduce the method of using GAN-generated texture image as the cover image and hide the secret information into it [12]. The input noise to the generative network is mapped to a small patch chosen from the reference image. The synthesized texture image together with the secret message are sent to the information hiding network to generate the final stego image.
Augmented reality extends virtual scenes into real world by a series hardware and software collaboration. It presents the virtual object into the real-world environment through a specific hardware. In order to realize the most realistic scene, the hardware and the augmented reality algorithm has to work seamlessly. Also, the interaction between the virtual objects and user should be just like in the real-world environment. Augmented reality can be utilized in many areas such as gaming, education, commercial marketing, advertisement, scientific research and military operations etc [13,14,15,16,17,18,19]. In the past few years, AR was mostly achieved through expensive head-mounted sets such as AR goggles, projection-based camera systems, and AR-extended professional devices such as x-ray scanners [13,14,16,18,20,21,22,23]. Generally, these devices work by transforming the real-world scene to a live video feed and integrating the video feed with the computer-generated objects. Moreover, with the rapid development of modern mobile devices such as iPhone and iPad, it seems that the most promising hardware to achieve this immersive user experience is through these commonly used handheld mobile devices. Gervautz and Schmalstieg investigate the developments in handheld AR and put forward several challenges on implementing augmented reality on mobile devices [18]. They address the challenges from three aspects including localization tracking, interaction, and visualization and summarize several available solutions. Carmigniani et al. survey the technology, systems and applications in Augmented Reality and point out the possible future directions for AR applications [16]. Four types of applications are summarized which include advertising and commercial, entertainment and education, medical, and mobile application for iPhones. One of the contributions of their survey is that they focused more on the mobile augmented reality applications and addressed the privacy and ethical issues along the way. Kroeker mentioned that current AR applications continue to be much more compact and powerful, and many of which require nothing more than a current-generation smartphone [17]. However, the complete augmented reality infrastructure standard is hard to build and the acceptance of the applications could also become a serious problem due to the privacy and security concerns [24]. Gattullo et al. propose a method which can support technical writers in the creation of technical documentation suitable for augmented reality interfaces and also point out that AR is one of the most promising technology for technical manuals in the context of Industry 4.0 [25]. Kohn and Harborth analyze the application of utilizing AR in the manufacturing industry and conclude that augmented reality will surely change the manufacturing industry but still needs further improvements to achieve the transformation [26]. Grubert et al. present a novel Pervasive Augmented Reality concept which expects to improve the user experience based on particular context [27]. With the announce of the ARKit by Apple company, augmented reality can be easily implemented and taken to more users. ARKit manages most of the device motion tracking callbacks, camera scene capture and image scene processing. With the even more powerful compute capability on mobile devices, machine learning or even deep learning method is also possible. By combining the modern deep learning computer vision algorithms with augmented reality, the real-world environment can be used as a whole new information carrier.
Zhou et al. first proposed the concept of coverless information hiding and addressed that coverless information hiding methods take advantage of the independent carrier itself to represent the secret message [28]. Coverless information hiding typically works by designing a carrier-secret matching policy in order to minimize the secret detection rate from third party tools [29]. Augmented reality based information hiding approach follows the general idea of coverless information hiding methods but further extends its flexibility.
The key of integrating augmented reality with information hiding lies in the secret mapping policy. The mapping policy follows the idea of representing one object in another form. For example, an image of the Statue of Liberty could actually represent the Mona Lisa portrait, a specific coordinate could actually be a bunch of hidden messages. Moreover, a specific image could combine a specific location information such as GPS and compass information to represent even more complicated things. The matched object here serves as the traditional information hiding key for deciphering the messages in this scheme. This kind of mapping strategy is simple and widely used at the early stage of information hiding. The augmented system detects the shared key objects and restores the message into the real-world environment. However, the biggest problem of this kind of mapping scheme lies in the limitation of the matching library. The matching database needs to be extended every time a new message is added which makes it less convenient and practical. Currently, some augmented reality frameworks do have the feature which allows multiple users to share one unique AR scene. This kind of feature provides a new scheme to transmit the secret messages between users but generally still requires the users to be near to each other.
As mentioned before, multiply secret keys can be combined in order to largely extend the secret matching library. Generally, more combination of keys would generate more mapping branches but also increase the complexity of the mapping library. In order to conquer the conundrum addressed by the simple key-secret mapping scheme, deep neural network based architecture can be a suitable solution. Deep neural network is good at mapping complex object features into different patterns and shows great robustness in various tasks. Deep neural network is usually trained with a specific dataset which contains the representative data of the corresponding task. In term of the secret matching scenario, the traditional matching policy can be implemented through the robust neural network and obtain the outcome in a general manner. Furthermore, various deep learning methods can be applied throughout the architecture in conjunction with the augmented reality. For instance, object detection methods can be applied to images so that the system could detect both the class and the location of certain objects existed in the image. The object detection rate can be used as an additional standard for secret matching such that the carrier is considered to be detected or the secret is allowed to be revealed only when the detection rate is above certain threshold. In order to enlarge the text library, natural language processing technique can be integrated in the secret revealing process for secret text restoration. Other advanced deep learning approaches such as image caption, image segmentation, and even image generation are also good candidates when it comes to image matching situation.
The proposed framework is shown in Figure 1 with the explanation and discussion down below. Two parties desire to transmit secret information between each other. Here A represents the message sender and B represents the secret receiver. A party wishes to hide secret messages inside another carrier and transmits the modified carrier to B side. The secret carrier should not look suspicious to another third party and the secret information should be able to be lossless restored when B extracts the information from the secret carrier. As Figure 1 shows, the secret messages are matched to different target images or different objects beforehand. Then the matched objects are sent to a secret hiding model for further process. Party A is in charge of the secret matching and the information hiding model is shared between two parties. For instance, suppose the secret message is matched to the real-world dog category, when A party tries to send message to B, he just needs to inform B what object he should look for in order to decode the secret message. The matched real-world dog category serves as the traditional information hiding key for deciphering the messages in this context. B can use the AR based information hiding system to find certain dog object and decode the secret message to the real world after receiving the secret key. Figure 2 depicts the mapping strategy in detail.
As Figure 2 shows, most object detection methods can detect certain object which belongs to specific class. The secret communication parties can transmit a concrete coordinate, a specific direction together with the target objects to enhance the secret communication. The secret variety and secrecy can be greatly enhanced by combining different elements together. The authorized receiver can point the augmented reality system to a specific direction and move to a specific location to decode the message.
The proposed mapping architecture can improve the secret embedding rate dramatically in potential since the target object can correspond to any kind of message. The secret communication parties usually share the secret messages via binary, text or image formats. This is mainly constrained by the secret embedding rate and detection rate. In general, by increasing the secret embedding rate, the secret detection rate would increase dramatically simultaneously. However, the augmented reality technique can expand the carrier into other formats while still maintaining high secret embedding rate. For instance, three-dimensional (3D) models and even videos can be transmitted as the original secret messages which would be a huge improvement compared to conventional information hiding methods. Figure 3 demonstrates a feasible framework which utilizes deep learning technique to realize three-dimensional model based information hiding via augmented reality platform. The video frame or picture is first segmented using object detection and object segmentation methods. Then the segmented object is further processed to achieve secret transmission. Three schemes are proposed regarding to the secret embedding process. The first scheme attempts to hide the secret message in the two-dimensional (2D) object beforehand and then transform the stego object into three-dimensional object. The second attempts to hide the secret message during the object transformation process by simultaneously learning an information hiding algorithm. The last scheme tries to embed the secret messages inside the transformed 3D object afterwards by modifying the 3D point cloud information. The transformation process is achieved by a neural network which learns a 3D point cloud generation algorithm that maps the 2D coordinates into 3D space [30,31,32,33,34,35,36,37,38].
A system prototype is developed according to the proposed architecture using iOS SDK. The system is designed under the image matching logic which corresponds the secret message to concrete real-world scene objects. It recognizes the specific image or object and then presents the corresponding secret message in the real-world environment. In this prototype system, the image to be recognized serves as the carrier of the secret message and is mapped to the designated secret message. The system can only be allowed to present the secret message when it detects the corresponding mapping image or object inside the real-world scene. Furthermore, the system is also implemented to show the secret message according to the specific object class and the detection rate of specific object class. The object detection method used in the system is implemented using You Only Look Once (YOLO)v3 which is a novel object detection method based on deep convolutional neural network [6]. YOLO can achieve a relative high detection accuracy while remaining a high detection speed which is suitable for real-time applications. The YOLO network used in this system is derived from the original network with some modifications in order to run on mobile devices and is trained using COCO2017 dataset which contains 80 categories of different objects. The programming language includes Swift, Python, C and C++, the YOLO model is trained using one GTX 1080 GPU and the system is tested on the Apple iPhone 7 plus.
The execution flow of the prototype system is shown in Figure 4. The secret matching policy is the combination of image recognition, object detection and orientation configuration. The communication parties share the target image, the desired object along with the detection threshold and certain orientation information. Figure 5 shows one scenario where the AR system works by first detecting a laptop object and a target image and is only allowed to reveal the secret message when the detection rate is above 80 percent and the target image is recognized. The secret message can also be matched to other objects and updated correspondingly.
This paper proposes a novel scheme which utilizes the augmented reality to achieve secret communication. The general idea for utilizing augmented reality to do secret communication lies in the secret matching policy which derives from coverless information hiding. The secret message is mapped to secret keys and can only be revealed when the corresponding key object is detected by the AR system. The mapping scheme can either derive from a concrete matching database or base on a certain deep learning model. The former only works when the mapping scheme is inside the database which is less flexible while the latter is more complicated to design but has fewer limitations. Several deep-learning based mapping schemes are mentioned in this paper including object detection, image caption, and natural language processing etc. The deep learning model is trained on the local server and the trained inference model is optimized to be deployed on the mobile device. The experimental AR system automatically detects the carrier and corresponds the secret messages to different objects. The secret message can only be revealed when the carrier object is detected with a certain confidence score. Clearly, information hiding combined with augmented reality is still in its infancy, the design of the secret mapping model is the crucial part of this kind of approach and is worthy of further study.
This work is supported by the National Key R & D Program of China under grant 2018YFB1003205; by the National Natural Science Foundation of China under grant U1836208, U1536206, U1836110, 61602253, 61672294; by the Jiangsu Basic Research Programs-Natural Science Foundation under grant numbers BK20181407; by the Priority Academic Program Development of Jiangsu Higher Education Institutions (PAPD) fund; by the Collaborative Innovation Center of Atmospheric Environment and Equipment Technology (CICAEET) fund, China.
The authors declare that there are no actual or potential conflicts of interest in relation to this article.
[1] | P. Pucer, Augmented reality,ZDR VESTN-Slov. Med. J.,80 (2011), 7–8. |
[2] | L. Avila and M. Bailey, Augment your reality,IEEE Comput. Graph.,36 (2016), 6–7. |
[3] | J. Wang, S. Lian and Y. Q. Shi, Hybrid multiplicative multi-watermarking in DWT domain,Multidim. Syst. Sign. P., 28 (2017), 617–636. |
[4] | Q. Cui, S. McIntosh and H. Sun, Identifying materials of photographic images and photorealistic computer generated graphics based on deep CNNs, CMC-Comput. Mater. Con., 55 (2018), 229–241. |
[5] | A. Krizhevsky, I. Sutskever and G. E. Hinton, Imagenet classification with deep convolutional neural networks, Adv. Neural Inform. Process. Syst., (2012), 1097–1105. |
[6] | J. Redmon, S. Divvala, R. Girshick, et al., You only look once: Unified, real-time object detection, Proceedings of the IEEE conference on computer vision and pattern recognition, (2016), 779–788. |
[7] | S. Ren, K. He, R. Girshick, et al., Faster R-CNN: Towards real-time object detection with region proposal networks, IEEE Trans. Pattern Anal. Mach. Intell, 6 (2017), 1137–1149. |
[8] | K. He, G. Gkioxari, P. Dollár, et al., Mask R-CNN,IEEE International Conference on Computer Vision,(2017), 2980–2988. |
[9] | C. Yuan, X. Li, Q. J. Wu, et al., Fingerprint liveness detection from different fingerprint materials using convolutional neural network and principal component analysis. Computers, CMC-Comput. Mater. Con., 53 (2017), 357–371. |
[10] | R. Rahim and M. S. Nadeem, End-to-end trained CNN encode-decoder networks for image steganography, preprint, arXiv:1711.07201. |
[11] | S. Baluja, Hiding images in plain sight: Deep steganography, Adv. Neural Inform. Process. Syst., (2017), 2066–2076. |
[12] | C. Li, Y. Jiang and M. Cheslyar, Embedding image through generated intermediate medium using deep convolutional generative adversarial network, CMC-Comput. Mater. Con.,56 (2018), 313–324. |
[13] | M. J. Kim, A framework for context immersion in mobile augmented reality,Automat. Constr.,33 (2013), 79–85. |
[14] | O. Bimber, What's real about augmented reality,Computer,45 (2012), 24–25. |
[15] | T. J. Brigham, Reality check: basics of augmented, virtual, and mixed reality,Med. Ref. Serv. Q.,36 (2017), 171–178. |
[16] | J. Carmigniani, B. Furht, M. Anisetti, et al., Augmented reality technologies, systems and applications,Multimed. Tools Appl.,51 (2011), 341–377. |
[17] | K. Kroeker, Mainstreaming augmented reality,Commun. ACM,53 (2010), 19–21. |
[18] | M. Gervautz and D. Schmalstieg, Anywhere interfaces using handheld augmented reality,Computer,45 (2012), 26–31. |
[19] | P. Diao and N. Shi, MARINS: A mobile smartphone AR system for pathfinding in a dark environment,Sensors,18 (2018), 3442.1–3442.14. |
[20] | J. S. Park, AR-Room: a rapid prototyping framework for augmented reality applications,Multimed. Tools Appl.,55 (2011), 725–746. |
[21] | M. R. Mine, J. Van Baar, A. Grundhofer, et al., Projection-based augmented reality in disney theme parks,Computer,45 (2012), 32–40. |
[22] | G. Papagiannakis, G. Singh and N. Magnenat-Thalmann, A survey of mobile and wireless technologies for augmented reality systems,Comput. Animat. Virt. W.,19 (2008), 3–22. |
[23] | T. Langlotz, D. Wagner and A. Mulloni, Online creation of panoramic augmented reality annotations on mobile phones,IEEE Pervas. Comput.,11 (2012), 56–63. |
[24] | F. Roesner, T. Kohno and D. Molnar, Security and privacy for augmented reality systems,Commun. ACM,57 (2014), 88–96. |
[25] | M. Gattullo, G. W. Scurati, M. Fiorentino, et al., Towards augmented reality manuals for industry 4.0: A methodology,Robot CIM-INT Manuf.,56 (2019), 276–286. |
[26] | V, Kohn and D. Harborth, Augmented Reality–A game changing technology for manufacturing process?, Twenty-Sixth European Conference on Information Systems, (2018). |
[27] | J. Grubert, T. Langlotz, S. Zollmann, et al., Towards pervasive augmented reality: Context-awareness in augmented reality,IEEE T. Vis. Comput. Gr.,23 (2017), 1706–1724. |
[28] | Z. Zhou, Q. M. J. Wu, C. Yang, et al., Coverless Image Steganography Using Histograms of Oriented Gradients-based Hashing Algorithm, J. Internet Technol., 18 (2017), 1177–1184. |
[29] | Y. Cao, Z. Zhou, C. Yang, et al., Dynamic content selection framework applied to coverless information hiding,J. Internet Technol.,19 (2018), 1179–1186. |
[30] | H. Fan, H. Su and L. Guibas, A point set generation network for 3D object reconstruction from a single image,2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), IEEE, (2017), 605–613. |
[31] | Z. Wu, S. Song, A. Khosla, et al., 3D shapenets: A deep representation for volumetric shapes, Proceedings of the IEEE conference on computer vision and pattern recognition, (2015), 1912–1920. |
[32] | M. Tatarchenko, A. Dosovitskiy and T. Brox, Multi-view 3d models from single images with a convolutional network,Knowl. Inf. Syst.,38 (2015), 231–257. |
[33] | M. Gadelha, S. Maji and R. Wang, 3D Shape Induction from 2D Views of Multiple Objects, 2017 International Conference on 3D Vision (3DV), (2017), 402–411. |
[34] | J. K. Pontes, C. Kong, S. Sridharan, et al, Image2mesh: A learning framework for single image 3d reconstruction, preprint, arXiv:1711.10669. |
[35] | P. Mandikal, N. KL and R. Venkatesh Babu, 3D-psrnet: Part segmented 3D point cloud reconstruction from a single image, Proceedings of the European Conference on Computer Vision (ECCV), (2018). |
[36] | A. Kar, S. Tulsiani, J. Carreira, et al., Category-specific object reconstruction from a single image, Proceedings of the IEEE conference on computer vision and pattern recognition, (2015), 1966–1974. |
[37] | A. Arsalan Soltani, H. Huang, J. Wu, et al, Synthesizing 3D Shapes via modeling multi-view depth maps and silhouettes with deep generative networks,2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), (2017), 1511–1519. |
[38] | A. X. Chang, T. Funkhouser, L. Guibas, et al., Shapenet: An information-rich 3d model repository, preprint, arXiv:1512.03012. |
1. | Guanghui Song, Hai Wang, Wei Wang, Artificial Intelligence-Assisted Fresco Restoration with Multiscale Line Drawing Generation, 2021, 2021, 1099-0526, 1, 10.1155/2021/5567966 | |
2. | Renyong Huang, Mingyi Sun, Network algorithm real-time depth image 3D human recognition for augmented reality, 2020, 1861-8200, 10.1007/s11554-020-01045-z | |
3. | Lingzhuang Meng, Lianshan Liu, Yanjun Peng, Xiaoli Wang, A data hiding scheme based on U-Net and wavelet transform, 2021, 09507051, 107022, 10.1016/j.knosys.2021.107022 | |
4. | Yalda Ghasemi, Heejin Jeong, Sung Ho Choi, Kyeong-Beom Park, Jae Yeol Lee, Deep learning-based object detection in augmented reality: A systematic review, 2022, 139, 01663615, 103661, 10.1016/j.compind.2022.103661 | |
5. | Yulan Ren, Yao Yang, Jiani Chen, Ying Zhou, Jiamei Li, Rui Xia, Yuan Yang, Qiao Wang, Xi Su, A scoping review of deep learning in cancer nursing combined with augmented reality: The era of intelligent nursing is coming, 2022, 9, 23475625, 100135, 10.1016/j.apjon.2022.100135 | |
6. | Maria Chatrasingh, Cholatip Wiratkapun, Jackrit Suthakorn, A generalized closed-form solution for 3D registration of two-point sets under isotropic and anisotropic scaling, 2023, 51, 22113797, 106746, 10.1016/j.rinp.2023.106746 | |
7. | Chia-Chen Lin, Aristophane Nshimiyimana, Morteza SaberiKamarposhti, Ersin Elbasi, Authentication Framework for Augmented Reality with Data-Hiding Technique, 2024, 16, 2073-8994, 1253, 10.3390/sym16101253 | |
8. | David Cortes, Belen Bermejo, Carlos Juiz, The use of CNNs in VR/AR/MR/XR: a systematic literature review, 2024, 28, 1434-9957, 10.1007/s10055-024-01044-6 | |
9. | Prabhash Kumar Singh, Biswapati Jana, Kakali Datta, Partha Chowdhuri, Pabitra Pal, 2023, Chapter 25, 978-981-19-5190-9, 303, 10.1007/978-981-19-5191-6_25 |