The uncanny valley hypothesis states that users may experience discomfort when interacting with almost human-like artificial characters. Advancements in artificial intelligence, robotics, and computer graphics have led to the development of life-like virtual humans and humanoid robots. Revisiting this hypothesis is necessary to check whether they positively or negatively affect the current population, who are highly accustomed to the latest technologies.
Methods
In this study, we present a unique evaluation of the uncanny valley hypothesis by allowing participants to interact live with four humanoid robots that have varying levels of human-likeness. Each participant completed a survey questionnaire to evaluate the affinity of each robot. Additionally, we used deep learning methods to quantify the participants’ emotional states using multimodal cues, including visual, audio, and text cues, by recording the participant–robot interactions.
Results
Multi-modal analysis and surveys provided interesting results and insights into the uncanny valley hypothesis.
{"title":"Uncanny valley for interactive social agents: an experimental study","authors":"Nidhi Mishra , Manoj Ramanathan , Gauri Tulsulkar , Nadia Magneat Thalmann","doi":"10.1016/j.vrih.2022.08.003","DOIUrl":"10.1016/j.vrih.2022.08.003","url":null,"abstract":"<div><h3>Background</h3><p>The uncanny valley hypothesis states that users may experience discomfort when interacting with almost human-like artificial characters. Advancements in artificial intelligence, robotics, and computer graphics have led to the development of life-like virtual humans and humanoid robots. Revisiting this hypothesis is necessary to check whether they positively or negatively affect the current population, who are highly accustomed to the latest technologies.</p></div><div><h3>Methods</h3><p>In this study, we present a unique evaluation of the uncanny valley hypothesis by allowing participants to interact live with four humanoid robots that have varying levels of human-likeness. Each participant completed a survey questionnaire to evaluate the affinity of each robot. Additionally, we used deep learning methods to quantify the participants’ emotional states using multimodal cues, including visual, audio, and text cues, by recording the participant–robot interactions.</p></div><div><h3>Results</h3><p>Multi-modal analysis and surveys provided interesting results and insights into the uncanny valley hypothesis.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"4 5","pages":"Pages 393-405"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S209657962200078X/pdf?md5=006cc0cfa178979a31eb04f193763508&pid=1-s2.0-S209657962200078X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125618019","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-01DOI: 10.1016/j.vrih.2022.08.007
Yicheng Zhao , Han Zhang , Ping Lu , Ping Li , Enhua Wu , Bin Sheng
Background
Exploring correspondences across multiview images is the basis of various computer vision tasks. However, most existing methods have limited accuracy under challenging conditions.
Method
To learn more robust and accurate correspondences, we propose DSD-MatchingNet for local feature matching in this study. First, we develop a deformable feature extraction module to obtain multilevel feature maps, which harvest contextual information from dynamic receptive fields. The dynamic receptive fields provided by the deformable convolution network ensure that our method obtains dense and robust correspondence. Second, we utilize sparse-to-dense matching with symmetry of correspondence to implement accurate pixel-level matching, which enables our method to produce more accurate correspondences.
Result
Experiments show that our proposed DSD-MatchingNet achieves a better performance on the image matching benchmark, as well as on the visual localization benchmark. Specifically, our method achieved 91.3% mean matching accuracy on the HPatches dataset and 99.3% visual localization recalls on the Aachen Day-Night dataset.
{"title":"DSD-MatchingNet: Deformable sparse-to-dense feature matching for learning accurate correspondences","authors":"Yicheng Zhao , Han Zhang , Ping Lu , Ping Li , Enhua Wu , Bin Sheng","doi":"10.1016/j.vrih.2022.08.007","DOIUrl":"10.1016/j.vrih.2022.08.007","url":null,"abstract":"<div><h3>Background</h3><p>Exploring correspondences across multiview images is the basis of various computer vision tasks. However, most existing methods have limited accuracy under challenging conditions.</p></div><div><h3>Method</h3><p>To learn more robust and accurate correspondences, we propose DSD-MatchingNet for local feature matching in this study. First, we develop a deformable feature extraction module to obtain multilevel feature maps, which harvest contextual information from dynamic receptive fields. The dynamic receptive fields provided by the deformable convolution network ensure that our method obtains dense and robust correspondence. Second, we utilize sparse-to-dense matching with symmetry of correspondence to implement accurate pixel-level matching, which enables our method to produce more accurate correspondences.</p></div><div><h3>Result</h3><p>Experiments show that our proposed DSD-MatchingNet achieves a better performance on the image matching benchmark, as well as on the visual localization benchmark. Specifically, our method achieved 91.3% mean matching accuracy on the HPatches dataset and 99.3% visual localization recalls on the Aachen Day-Night dataset.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"4 5","pages":"Pages 432-443"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2096579622000821/pdf?md5=b3b9d92de1f1714de8cb8ab71d43808f&pid=1-s2.0-S2096579622000821-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133210436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-01DOI: 10.1016/j.vrih.2022.08.006
Philipp Braun, Michaela Grafelmann, Felix Gill, Hauke Stolz, Johannes Hinckeldeyn, Ann-Kathrin Lange
Background
Virtual reality (VR) applications can be used to provide comprehensive training scenarios that are difficult or impossible to represent in physical configurations. This includes team training for emergency services such as firefighting. Creating a high level of immersion is essential for achieving effective virtual training. In this respect, motion-capture systems offer the possibility of creating highly immersive multi-user training experiences, including full-body avatars.
Methods
This study presents a preliminary prototype that helps extinguish a virtual fire on a container ship as a VR training scenario. The prototype provides a full-body and multi-user VR experience based on the synthesis of position data provided by the motion-capture system and orientation data from the VR headsets. Moreover, the prototype facilitates an initial evaluation of the results.
Results
The results confirm the value of using VR for training procedures that are difficult to implement in the real world. Furthermore, the results show that motion-capture-based VR technologies are particularly useful for firefighting training, in which participants can collaborate in difficult-to-access environments. However, this study also indicates that increasing the immersion in such training remains a challenge.
Conclusions
This study presents a prototypical VR application that enables the multi-user training of maritime firefighters. Future research should evaluate the initial results, provide more extensive training scenarios, and measure the training progress.
{"title":"Virtual reality for immersive multi-user firefighter-training scenarios","authors":"Philipp Braun, Michaela Grafelmann, Felix Gill, Hauke Stolz, Johannes Hinckeldeyn, Ann-Kathrin Lange","doi":"10.1016/j.vrih.2022.08.006","DOIUrl":"10.1016/j.vrih.2022.08.006","url":null,"abstract":"<div><h3>Background</h3><p>Virtual reality (VR) applications can be used to provide comprehensive training scenarios that are difficult or impossible to represent in physical configurations. This includes team training for emergency services such as firefighting. Creating a high level of immersion is essential for achieving effective virtual training. In this respect, motion-capture systems offer the possibility of creating highly immersive multi-user training experiences, including full-body avatars.</p></div><div><h3>Methods</h3><p>This study presents a preliminary prototype that helps extinguish a virtual fire on a container ship as a VR training scenario. The prototype provides a full-body and multi-user VR experience based on the synthesis of position data provided by the motion-capture system and orientation data from the VR headsets. Moreover, the prototype facilitates an initial evaluation of the results.</p></div><div><h3>Results</h3><p>The results confirm the value of using VR for training procedures that are difficult to implement in the real world. Furthermore, the results show that motion-capture-based VR technologies are particularly useful for firefighting training, in which participants can collaborate in difficult-to-access environments. However, this study also indicates that increasing the immersion in such training remains a challenge.</p></div><div><h3>Conclusions</h3><p>This study presents a prototypical VR application that enables the multi-user training of maritime firefighters. Future research should evaluate the initial results, provide more extensive training scenarios, and measure the training progress.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"4 5","pages":"Pages 406-417"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S209657962200081X/pdf?md5=1287c6b99ffe058108e336cc8bf7aca8&pid=1-s2.0-S209657962200081X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131573670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-01DOI: 10.1016/j.vrih.2022.08.008
Zhuo Shi , Mingrui Li , Meng Wang , Jing Shen , Wei Chen , Xiaonan Luo
Background
Data-driven event analysis has gradually become the backbone of modern competitive sports analysis. Competitive sports data analysis tasks increasingly use computer vision and machine-learning models for intelligent data analysis. Existing sports visualization systems focus on the player–team data visualization, which is not intuitive enough for team season win–loss data and game time-series data visualization and neglects the prediction of all-star players.
Methods
This study used an interactive visualization system designed with parallel aggregated ordered hypergraph dynamic hypergraphs, Calliope visualization data story technology, and iStoryline narrative visualization technology to visualize the regular statistics and game time data of players and teams. NPIPVis includes dynamic hypergraphs of a teamʹs wins and losses and game plot narrative visualization components. In addition, an integrated learning-based all-star player prediction model, SRR-voting, which starts from the existing minority and majority samples, was proposed using the synthetic minority oversampling technique and RandomUnderSampler methods to generate and eliminate samples of a certain size to balance the number of allstar and average players in the datasets. Next, a random forest algorithm was introduced to extract and construct the features of players and combined with the voting integrated model to predict the all-star players, using Grid- SearchCV, to optimize the hyperparameters of each model in integrated learning and then combined with five-fold cross-validation to improve the generalization ability of the model. Finally, the SHapley Additive exPlanations (SHAP) model was introduced to enhance the interpretability of the model.
Results
The experimental results of comparing the SRR-voting model with six common models show that the accuracy, F1-score, and recall metrics are significantly improved, which verifies the effectiveness and practicality of the SRR-voting model.
Conclusions
This study combines data visualization and machine learning to design a National Basketball Association data visualization system to help the general audience visualize game data and predict all-star players; this can also be extended to other sports events or related fields.
{"title":"NPIPVis: A visualization system involving NBA visual analysis and integrated learning model prediction","authors":"Zhuo Shi , Mingrui Li , Meng Wang , Jing Shen , Wei Chen , Xiaonan Luo","doi":"10.1016/j.vrih.2022.08.008","DOIUrl":"10.1016/j.vrih.2022.08.008","url":null,"abstract":"<div><h3>Background</h3><p>Data-driven event analysis has gradually become the backbone of modern competitive sports analysis. Competitive sports data analysis tasks increasingly use computer vision and machine-learning models for intelligent data analysis. Existing sports visualization systems focus on the player–team data visualization, which is not intuitive enough for team season win–loss data and game time-series data visualization and neglects the prediction of all-star players.</p></div><div><h3>Methods</h3><p>This study used an interactive visualization system designed with parallel aggregated ordered hypergraph dynamic hypergraphs, Calliope visualization data story technology, and iStoryline narrative visualization technology to visualize the regular statistics and game time data of players and teams. NPIPVis includes dynamic hypergraphs of a teamʹs wins and losses and game plot narrative visualization components. In addition, an integrated learning-based all-star player prediction model, SRR-voting, which starts from the existing minority and majority samples, was proposed using the synthetic minority oversampling technique and RandomUnderSampler methods to generate and eliminate samples of a certain size to balance the number of allstar and average players in the datasets. Next, a random forest algorithm was introduced to extract and construct the features of players and combined with the voting integrated model to predict the all-star players, using Grid- SearchCV, to optimize the hyperparameters of each model in integrated learning and then combined with five-fold cross-validation to improve the generalization ability of the model. Finally, the SHapley Additive exPlanations (SHAP) model was introduced to enhance the interpretability of the model.</p></div><div><h3>Results</h3><p>The experimental results of comparing the SRR-voting model with six common models show that the accuracy, F1-score, and recall metrics are significantly improved, which verifies the effectiveness and practicality of the SRR-voting model.</p></div><div><h3>Conclusions</h3><p>This study combines data visualization and machine learning to design a National Basketball Association data visualization system to help the general audience visualize game data and predict all-star players; this can also be extended to other sports events or related fields.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"4 5","pages":"Pages 444-458"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2096579622000833/pdf?md5=1a85324d30377a749ed5c9c70fb6f227&pid=1-s2.0-S2096579622000833-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125618325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-01DOI: 10.1016/j.vrih.2022.08.004
Lesley Istead , Joe Istead , Andreea Pocol , Craig S. Kaplan
Background
Gesture drawing is a type of fluid, fast sketch with loose and roughly drawn lines that captures the motion and feeling of a subject. Although style transfer methods, which are able to learn a style from an input image and apply it to a secondary image, can reproduce many styles, they are currently unable to produce the flowing strokes of gesture drawings.
Method
In this paper, we present a method for producing gesture drawings that roughly depict objects or scenes with loose dancing contours and frantic textures. By following a gradient field, our method adapts stroke-based painterly rendering algorithms to produce long curved strokes. A rough, overdrawn appearance is created through a progressive refinement. In addition, we produce rough hatch strokes by altering the stroke direction. These add optional shading to gesture drawings.
Results
A wealth of parameters provide users the ability to adjust the output style, from short and rapid strokes to long and fluid strokes, and from swirling to straight lines. Potential stylistic outputs include pen-and-ink and colored pencil. We present several generated gesture drawings and discuss the application of our method to video.
Conclusion
Our stroke-based rendering algorithm produces convincing gesture drawings with numerous controllable parameters, permitting the creation of a variety of styles.
{"title":"A simple, stroke-based method for gesture drawing","authors":"Lesley Istead , Joe Istead , Andreea Pocol , Craig S. Kaplan","doi":"10.1016/j.vrih.2022.08.004","DOIUrl":"10.1016/j.vrih.2022.08.004","url":null,"abstract":"<div><h3>Background</h3><p>Gesture drawing is a type of fluid, fast sketch with loose and roughly drawn lines that captures the motion and feeling of a subject. Although style transfer methods, which are able to learn a style from an input image and apply it to a secondary image, can reproduce many styles, they are currently unable to produce the flowing strokes of gesture drawings.</p></div><div><h3>Method</h3><p>In this paper, we present a method for producing gesture drawings that roughly depict objects or scenes with loose dancing contours and frantic textures. By following a gradient field, our method adapts stroke-based painterly rendering algorithms to produce long curved strokes. A rough, overdrawn appearance is created through a progressive refinement. In addition, we produce rough hatch strokes by altering the stroke direction. These add optional shading to gesture drawings.</p></div><div><h3>Results</h3><p>A wealth of parameters provide users the ability to adjust the output style, from short and rapid strokes to long and fluid strokes, and from swirling to straight lines. Potential stylistic outputs include pen-and-ink and colored pencil. We present several generated gesture drawings and discuss the application of our method to video.</p></div><div><h3>Conclusion</h3><p>Our stroke-based rendering algorithm produces convincing gesture drawings with numerous controllable parameters, permitting the creation of a variety of styles.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"4 5","pages":"Pages 381-392"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2096579622000791/pdf?md5=5d6ede6955247fdfc333f73a0cddaa0d&pid=1-s2.0-S2096579622000791-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122622709","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-10-01DOI: 10.1016/j.vrih.2022.08.005
Chuxuan Li , Ran Yi , Saba Ghazanfar Ali , Lizhuang Ma , Enhua Wu , Jihong Wang , Lijuan Mao , Bin Sheng
Background
Monocular depth estimation aims to predict a dense depth map from a single RGB image, and has important applications in 3D reconstruction, automatic driving, and augmented reality. However, existing methods directly feed the original RGB image into the model to extract depth features without avoiding the interference of depth-irrelevant information on depth-estimation accuracy, which leads to inferior performance.
Methods
To remove the influence of depth-irrelevant information and improve the depth-prediction accuracy, we propose RADepthNet, a novel reflectance-guided network that fuses boundary features. Specifically, our method predicts depth maps using the following three steps: (1) Intrinsic Image Decomposition. We propose a reflectance extraction module consisting of an encoder-decoder structure to extract the depth-related reflectance. Through an ablation study, we demonstrate that the module can reduce the influence of illumination on depth estimation. (2) Boundary Detection. A boundary extraction module, consisting of an encoder, refinement block, and upsample block, was proposed to better predict the depth at object boundaries utilizing gradient constraints. (3) Depth Prediction Module. We use an encoder different from (2) to obtain depth features from the reflectance map and fuse boundary features to predict depth. In addition, we proposed FIFADataset, a depth-estimation dataset applied in soccer scenarios.
Results
Extensive experiments on a public dataset and our proposed FIFADataset show that our method achieves state-of-the-art performance.
{"title":"RADepthNet: Reflectance-aware monocular depth estimation","authors":"Chuxuan Li , Ran Yi , Saba Ghazanfar Ali , Lizhuang Ma , Enhua Wu , Jihong Wang , Lijuan Mao , Bin Sheng","doi":"10.1016/j.vrih.2022.08.005","DOIUrl":"10.1016/j.vrih.2022.08.005","url":null,"abstract":"<div><h3>Background</h3><p>Monocular depth estimation aims to predict a dense depth map from a single RGB image, and has important applications in 3D reconstruction, automatic driving, and augmented reality. However, existing methods directly feed the original RGB image into the model to extract depth features without avoiding the interference of depth-irrelevant information on depth-estimation accuracy, which leads to inferior performance.</p></div><div><h3>Methods</h3><p>To remove the influence of depth-irrelevant information and improve the depth-prediction accuracy, we propose RADepthNet, a novel reflectance-guided network that fuses boundary features. Specifically, our method predicts depth maps using the following three steps: (1) Intrinsic Image Decomposition. We propose a reflectance extraction module consisting of an encoder-decoder structure to extract the depth-related reflectance. Through an ablation study, we demonstrate that the module can reduce the influence of illumination on depth estimation. (2) Boundary Detection. A boundary extraction module, consisting of an encoder, refinement block, and upsample block, was proposed to better predict the depth at object boundaries utilizing gradient constraints. (3) Depth Prediction Module<strong>.</strong> We use an encoder different from (2) to obtain depth features from the reflectance map and fuse boundary features to predict depth. In addition, we proposed FIFADataset, a depth-estimation dataset applied in soccer scenarios.</p></div><div><h3>Results</h3><p>Extensive experiments on a public dataset and our proposed FIFADataset show that our method achieves state-of-the-art performance.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"4 5","pages":"Pages 418-431"},"PeriodicalIF":0.0,"publicationDate":"2022-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2096579622000808/pdf?md5=fc1d9cddf0180762f5b3a461f1d2e01d&pid=1-s2.0-S2096579622000808-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116232217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-08-01DOI: 10.1016/j.vrih.2022.03.002
Imran Ahmed , Misbah Ahmad , Gwanggil Jeon
Background
Digital twins are virtual representations of devices and processes that capture the physical properties of the environment and operational algorithms/techniques in the context of medical devices and technologies. Digital twins may allow healthcare organizations to determine methods of improving medical processes, enhancing patient experience, lowering operating expenses, and extending the value of care. During the present COVID-19 pandemic, various medical devices, such as X-rays and CT scan machines and processes, are constantly being used to collect and analyze medical images. When collecting and processing an extensive volume of data in the form of images, machines and processes sometimes suffer from system failures, creating critical issues for hospitals and patients.
Methods
To address this, we introduce a digital-twin-based smart healthcare system integrated with medical devices to collect information regarding the current health condition, configuration, and maintenance history of the device/machine/system. Furthermore, medical images, that is, X-rays, are analyzed by using a deep-learning model to detect the infection of COVID-19. The designed system is based on the cascade recurrent convolution neural network (RCNN) architecture. In this architecture, the detector stages are deeper and more sequentially selective against small and close false positives. This architecture is a multi-stage extension of the RCNN model and sequentially trained using the output of one stage for training the other. At each stage, the bounding boxes are adjusted to locate a suitable value of the nearest false positives during the training of the different stages. In this manner, the arrangement of detectors is adjusted to increase the intersection over union, overcoming the problem of overfitting. We train the model by using X-ray images as the model was previously trained on another dataset.
Results
The developed system achieves good accuracy during the detection phase of COVID-19. The experimental outcomes reveal the efficiency of the detection architecture, which yields a mean average precision rate of 0.94.
{"title":"Integrating digital twins and deep learning for medical image analysis in the era of COVID-19","authors":"Imran Ahmed , Misbah Ahmad , Gwanggil Jeon","doi":"10.1016/j.vrih.2022.03.002","DOIUrl":"10.1016/j.vrih.2022.03.002","url":null,"abstract":"<div><h3>Background</h3><p>Digital twins are virtual representations of devices and processes that capture the physical properties of the environment and operational algorithms/techniques in the context of medical devices and technologies. Digital twins may allow healthcare organizations to determine methods of improving medical processes, enhancing patient experience, lowering operating expenses, and extending the value of care. During the present COVID-19 pandemic, various medical devices, such as X-rays and CT scan machines and processes, are constantly being used to collect and analyze medical images. When collecting and processing an extensive volume of data in the form of images, machines and processes sometimes suffer from system failures, creating critical issues for hospitals and patients.</p></div><div><h3>Methods</h3><p>To address this, we introduce a digital-twin-based smart healthcare system integrated with medical devices to collect information regarding the current health condition, configuration, and maintenance history of the device/machine/system. Furthermore, medical images, that is, X-rays, are analyzed by using a deep-learning model to detect the infection of COVID-19. The designed system is based on the cascade recurrent convolution neural network (RCNN) architecture. In this architecture, the detector stages are deeper and more sequentially selective against small and close false positives. This architecture is a multi-stage extension of the RCNN model and sequentially trained using the output of one stage for training the other. At each stage, the bounding boxes are adjusted to locate a suitable value of the nearest false positives during the training of the different stages. In this manner, the arrangement of detectors is adjusted to increase the intersection over union, overcoming the problem of overfitting. We train the model by using X-ray images as the model was previously trained on another dataset.</p></div><div><h3>Results</h3><p>The developed system achieves good accuracy during the detection phase of COVID-19. The experimental outcomes reveal the efficiency of the detection architecture, which yields a mean average precision rate of 0.94.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"4 4","pages":"Pages 292-305"},"PeriodicalIF":0.0,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2096579622000183/pdf?md5=1f5a53060a043bec60b5fd3de876ef4d&pid=1-s2.0-S2096579622000183-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42600248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-08-01DOI: 10.1016/j.vrih.2022.05.003
Christos L. Stergiou, Kostas E. Psannis
This work surveys and illustrates multiple open challenges in the field of industrial Internet of Things (IoT)-based big data management and analysis in cloud environments. Challenges arising from the fields of machine learning in cloud infrastructures, artificial intelligence techniques for big data analytics in cloud environments, and federated learning cloud systems are elucidated. Additionally, reinforcement learning, which is a novel technique that allows large cloud-based data centers, to allocate more energy-efficient resources is examined. Moreover, we propose an architecture that attempts to combine the features offered by several cloud providers to achieve an energy-efficient industrial IoT-based big data management framework (EEIBDM) established outside of every user in the cloud. IoT data can be integrated with techniques such as reinforcement and federated learning to achieve a digital twin scenario for the virtual representation of industrial IoT-based big data of machines and room temperatures. Furthermore, we propose an algorithm for determining the energy consumption of the infrastructure by evaluating the EEIBDM framework. Finally, future directions for the expansion of this research are discussed.
{"title":"Digital twin intelligent system for industrial internet of things-based big data management and analysis in cloud environments","authors":"Christos L. Stergiou, Kostas E. Psannis","doi":"10.1016/j.vrih.2022.05.003","DOIUrl":"10.1016/j.vrih.2022.05.003","url":null,"abstract":"<div><p>This work surveys and illustrates multiple open challenges in the field of industrial Internet of Things (IoT)-based big data management and analysis in cloud environments. Challenges arising from the fields of machine learning in cloud infrastructures, artificial intelligence techniques for big data analytics in cloud environments, and federated learning cloud systems are elucidated. Additionally, reinforcement learning, which is a novel technique that allows large cloud-based data centers, to allocate more energy-efficient resources is examined. Moreover, we propose an architecture that attempts to combine the features offered by several cloud providers to achieve an energy-efficient industrial IoT-based big data management framework (EEIBDM) established outside of every user in the cloud. IoT data can be integrated with techniques such as reinforcement and federated learning to achieve a digital twin scenario for the virtual representation of industrial IoT-based big data of machines and room temperatures. Furthermore, we propose an algorithm for determining the energy consumption of the infrastructure by evaluating the EEIBDM framework. Finally, future directions for the expansion of this research are discussed.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"4 4","pages":"Pages 279-291"},"PeriodicalIF":0.0,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2096579622000444/pdf?md5=77ac7ba395219ea4a1f3583a51767386&pid=1-s2.0-S2096579622000444-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132922825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-08-01DOI: 10.1016/j.vrih.2022.03.001
Marc Baaden
Background
Digital twins offer rich potential for exploration in virtual reality (VR). Using interactive molecular simulation approaches, they enable a human operator to access the physical properties of molecular objects and to build, manipulate, and study their assemblies. Integrative modeling and drug design are important applications of this technology.
Methods
In this study, head-mounted virtual reality displays connected to molecular simulation engines were used to create interactive and immersive digital twins. They were used to perform tasks relevant to specific use cases.
Results
Three areas were investigated, including model building, rational design, and tangible models. Here, we report several membrane-embedded systems of ion channels, viral components, and artificial water channels. We were able to improve and create molecular designs based on digital twins.
Conclusions
The molecular application domain offers great opportunities, and most of the technical and technological aspects have been solved. Wider adoption is expected once the onboarding of VR is simplified and the technology gains wider acceptance.
{"title":"Deep inside molecules — digital twins at the nanoscale","authors":"Marc Baaden","doi":"10.1016/j.vrih.2022.03.001","DOIUrl":"10.1016/j.vrih.2022.03.001","url":null,"abstract":"<div><h3>Background</h3><p>Digital twins offer rich potential for exploration in virtual reality (VR). Using interactive molecular simulation approaches, they enable a human operator to access the physical properties of molecular objects and to build, manipulate, and study their assemblies. Integrative modeling and drug design are important applications of this technology.</p></div><div><h3>Methods</h3><p>In this study, head-mounted virtual reality displays connected to molecular simulation engines were used to create interactive and immersive digital twins. They were used to perform tasks relevant to specific use cases.</p></div><div><h3>Results</h3><p>Three areas were investigated, including model building, rational design, and tangible models. Here, we report several membrane-embedded systems of ion channels, viral components, and artificial water channels. We were able to improve and create molecular designs based on digital twins.</p></div><div><h3>Conclusions</h3><p>The molecular application domain offers great opportunities, and most of the technical and technological aspects have been solved. Wider adoption is expected once the onboarding of VR is simplified and the technology gains wider acceptance.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"4 4","pages":"Pages 324-341"},"PeriodicalIF":0.0,"publicationDate":"2022-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2096579622000171/pdf?md5=c3f874da70ddd62619d89326c3770de9&pid=1-s2.0-S2096579622000171-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132137540","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}