首页 > 最新文献

Virtual Reality Intelligent Hardware最新文献

英文 中文
A data-based real-time petrochemical gas diffusion simulation approach on virtual reality 基于数据的虚拟现实石油化工气体扩散实时模拟方法
Q1 Computer Science Pub Date : 2023-06-01 DOI: 10.1016/j.vrih.2023.01.001
Min Yang , Yong Han , Chang Su , Xue Li

Background

Petrochemical products are flammable, explosive, and toxic, petrochemical accidents are generally extremely destructive. Therefore, disaster analysis and prediction and real-time simulation have become important means to control and reduce accident hazards.

Methods

In this study, a complete real-time simulation solution of gas diffusion with coordinate data and concentration data is proposed, which is mainly aimed at the simulation of the types of harmful gas leakage and diffusion accidents in the petrochemical industry. The rendering effect is more continuous and accurate through grid homogenization and trilinear interpolation. A data processing and rendering parallelization process is presented to improve simulation efficiency. Combines gas concentration and fragment transparency to synthesize transparent pixels in a scene. To ensure the approximate accuracy of the rendering effect, improve the efficiency of real-time rendering, and meet the requirement of intuitive perception using concentration data, a weighted blended order-independent transparency with enhanced alpha weight is presented, which can provide a more intuitive perception of hierarchical information of concentration data while preserving depth information. In this study, three order-independent transparency algorithms, depth peeling algorithm, weighted blended order-independent transparency and weighted blended order-independent transparency with enhanced alpha weight, are compared and analyzed for rendering image quality, rendering time, required memory, hierarchical information and so on.

Results

Using weighted blended order-independent transparency with enhanced alpha weight technique, the rendering time is shortened by 53.2% compared with the depth peeling algorithm, and the texture memory required is much smaller than the depth peeling algorithm. The rendering results of weighted blended order-independent transparency with enhanced alpha weight are approximately accurate compared with the depth peeling algorithm as ground truth, and there is no popping when surfaces pass through one another. At the same time, compared with weighted blended order-independent transparency, weighted blended OIT with enhanced alpha weight achieves an intuitive perception of hierarchical information of concentration data.

背景石油化工产品具有易燃、易爆、有毒等特点,石油化工事故一般具有极强的破坏性。因此,灾害分析预测和实时仿真已成为控制和减少事故危害的重要手段。方法本研究提出了一个完整的具有坐标数据和浓度数据的气体扩散实时模拟解决方案,主要针对石化行业有害气体泄漏和扩散事故的类型进行模拟。通过网格均匀化和三线性插值,渲染效果更加连续和准确。为了提高仿真效率,提出了一种数据处理和绘制并行化过程。将气体浓度和碎片透明度相结合,以合成场景中的透明像素。为了确保渲染效果的近似精度,提高实时渲染的效率,并满足使用浓度数据进行直观感知的要求,提出了一种具有增强阿尔法权重的加权混合顺序无关透明度,其可以在保留深度信息的同时提供对浓度数据的分层信息的更直观的感知。在本研究中,对深度剥离算法、加权混合顺序无关透明度和增强阿尔法权重的加权混合顺序独立透明度三种顺序无关透明度算法的渲染图像质量、渲染时间、所需内存进行了比较和分析,结果使用加权混合顺序无关透明度和增强阿尔法权重技术,与深度剥离算法相比,渲染时间缩短了53.2%,所需的纹理内存远小于深度剥离算法。与深度剥离算法相比,具有增强alpha权重的加权混合顺序无关透明度的渲染结果大致准确,并且当曲面相互穿过时不会出现爆裂。同时,与加权混合顺序无关透明度相比,具有增强阿尔法权重的加权混合OIT实现了对浓度数据层次信息的直观感知。
{"title":"A data-based real-time petrochemical gas diffusion simulation approach on virtual reality","authors":"Min Yang ,&nbsp;Yong Han ,&nbsp;Chang Su ,&nbsp;Xue Li","doi":"10.1016/j.vrih.2023.01.001","DOIUrl":"https://doi.org/10.1016/j.vrih.2023.01.001","url":null,"abstract":"<div><h3>Background</h3><p>Petrochemical products are flammable, explosive, and toxic, petrochemical accidents are generally extremely destructive. Therefore, disaster analysis and prediction and real-time simulation have become important means to control and reduce accident hazards.</p></div><div><h3>Methods</h3><p>In this study, a complete real-time simulation solution of gas diffusion with coordinate data and concentration data is proposed, which is mainly aimed at the simulation of the types of harmful gas leakage and diffusion accidents in the petrochemical industry. The rendering effect is more continuous and accurate through grid homogenization and trilinear interpolation. A data processing and rendering parallelization process is presented to improve simulation efficiency. Combines gas concentration and fragment transparency to synthesize transparent pixels in a scene. To ensure the approximate accuracy of the rendering effect, improve the efficiency of real-time rendering, and meet the requirement of intuitive perception using concentration data, a weighted blended order-independent transparency with enhanced alpha weight is presented, which can provide a more intuitive perception of hierarchical information of concentration data while preserving depth information. In this study, three order-independent transparency algorithms, depth peeling algorithm, weighted blended order-independent transparency and weighted blended order-independent transparency with enhanced alpha weight, are compared and analyzed for rendering image quality, rendering time, required memory, hierarchical information and so on.</p></div><div><h3>Results</h3><p>Using weighted blended order-independent transparency with enhanced alpha weight technique, the rendering time is shortened by 53.2% compared with the depth peeling algorithm, and the texture memory required is much smaller than the depth peeling algorithm. The rendering results of weighted blended order-independent transparency with enhanced alpha weight are approximately accurate compared with the depth peeling algorithm as ground truth, and there is no popping when surfaces pass through one another. At the same time, compared with weighted blended order-independent transparency, weighted blended OIT with enhanced alpha weight achieves an intuitive perception of hierarchical information of concentration data.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 3","pages":"Pages 266-278"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49833537","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
The Improvement of Iterative Closest Point with Edges of Projected Image 投影图像边缘迭代最近点的改进
Q1 Computer Science Pub Date : 2023-06-01 DOI: 10.1016/j.vrih.2022.09.001
Chen Wang

Background

There are many regular-shape objects in the artificial environment. It is difficult to distinguish the poses of these objects, when only geometric information is utilized. With the development of sensor technologies, we can utilize other information to solve this problem.

Methods

We propose an algorithm to register point clouds by integrating color information. The key idea of the algorithm is that we jointly optimize dense term and edge term. The dense term is built similarly to iterative closest point algorithm. In order to build the edge term, we extract the edges of the images obtained by projecting the point clouds. The edge term prevents the point clouds from sliding in registration. We utilize this loosely coupled method to fuse geometric and color information.

Results

The experiments demonstrate that edge image approach improves the precision and the algorithm is robust.

背景在人造环境中有许多规则形状的物体。当只利用几何信息时,很难区分这些物体的姿态。随着传感器技术的发展,我们可以利用其他信息来解决这个问题。方法提出了一种基于颜色信息的点云配准算法。该算法的核心思想是对密集项和边缘项进行联合优化。密集项的建立类似于迭代最接近点算法。为了建立边缘项,我们提取了通过投影点云获得的图像的边缘。边缘项可防止点云在配准中滑动。我们利用这种松散耦合的方法来融合几何和颜色信息。结果实验表明,边缘图像方法提高了算法的精度,算法具有较强的鲁棒性。
{"title":"The Improvement of Iterative Closest Point with Edges of Projected Image","authors":"Chen Wang","doi":"10.1016/j.vrih.2022.09.001","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.09.001","url":null,"abstract":"<div><h3>Background</h3><p>There are many regular-shape objects in the artificial environment. It is difficult to distinguish the poses of these objects, when only geometric information is utilized. With the development of sensor technologies, we can utilize other information to solve this problem.</p></div><div><h3>Methods</h3><p>We propose an algorithm to register point clouds by integrating color information. The key idea of the algorithm is that we jointly optimize dense term and edge term. The dense term is built similarly to iterative closest point algorithm. In order to build the edge term, we extract the edges of the images obtained by projecting the point clouds. The edge term prevents the point clouds from sliding in registration. We utilize this loosely coupled method to fuse geometric and color information.</p></div><div><h3>Results</h3><p>The experiments demonstrate that edge image approach improves the precision and the algorithm is robust.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 3","pages":"Pages 279-291"},"PeriodicalIF":0.0,"publicationDate":"2023-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49833536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Designing interactive glazing through an engineering psychology approach: Six augmented reality scenarios that envision future car human-machine interface 通过工程心理学方法设计交互式玻璃:六个增强现实场景,设想未来汽车人机界面
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.07.004
Wei Liu , Yancong Zhu , Ruonan Huang , Takumi Ohashi , Jan Auernhammer , Xiaonan Zhang , Ce Shi , Lu Wang

With more and more vehicles becoming autonomous, intelligent, and connected, paying attention to the future usage of car human-machine interface (HMI) with these vehicles should also get more relevant. While car HMI has been addressed in several scientific studies, little attention is being paid to designing and implementing interactive glazing into everyday (autonomous) driving contexts. Through reflecting on what was found before in theory and practice, we describe an engineering psychology practice and the design of six novel future user scenarios, which envision the application of a specific set of augmented reality (AR) support user interactions. We also present evaluations conducted with the scenarios and experiential prototypes and found that these AR scenarios support our target user groups in experiencing a new type of interactions. The overall evaluation was positive, with some valuable assessment results and suggestions. We envision that this paper will interest applied psychology educators who aspire to teach how to operationalize AR in a human-centered design (HCD) process to students with little preexisting expertise or little scientific knowledge about engineering psychology.

随着越来越多的车辆变得自动、智能和互联,关注未来汽车人机界面(HMI)与这些车辆的使用也应该变得更加相关。虽然汽车人机界面已经在一些科学研究中得到了解决,但很少有人关注在日常(自动)驾驶环境中设计和实现交互式玻璃。通过反思之前在理论和实践中发现的东西,我们描述了一个工程心理学实践和六个新的未来用户场景的设计,这些场景设想了一套特定的增强现实(AR)支持用户交互的应用。我们还对场景和体验原型进行了评估,发现这些AR场景支持我们的目标用户群体体验一种新型的交互。总体评价是积极的,并提出了一些有价值的评价结果和建议。我们设想,这篇论文将引起应用心理学教育者的兴趣,他们渴望向缺乏工程心理学专业知识或科学知识的学生教授如何在以人为本的设计(HCD)过程中操作AR。
{"title":"Designing interactive glazing through an engineering psychology approach: Six augmented reality scenarios that envision future car human-machine interface","authors":"Wei Liu ,&nbsp;Yancong Zhu ,&nbsp;Ruonan Huang ,&nbsp;Takumi Ohashi ,&nbsp;Jan Auernhammer ,&nbsp;Xiaonan Zhang ,&nbsp;Ce Shi ,&nbsp;Lu Wang","doi":"10.1016/j.vrih.2022.07.004","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.07.004","url":null,"abstract":"<div><p>With more and more vehicles becoming autonomous, intelligent, and connected, paying attention to the future usage of car human-machine interface (HMI) with these vehicles should also get more relevant. While car HMI has been addressed in several scientific studies, little attention is being paid to designing and implementing interactive glazing into everyday (autonomous) driving contexts. Through reflecting on what was found before in theory and practice, we describe an engineering psychology practice and the design of six novel future user scenarios, which envision the application of a specific set of augmented reality (AR) support user interactions. We also present evaluations conducted with the scenarios and experiential prototypes and found that these AR scenarios support our target user groups in experiencing a new type of interactions. The overall evaluation was positive, with some valuable assessment results and suggestions. We envision that this paper will interest applied psychology educators who aspire to teach how to operationalize AR in a human-centered design (HCD) process to students with little preexisting expertise or little scientific knowledge about engineering psychology.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 157-170"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49891616","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Outliers rejection in similar image matching 相似图像匹配中的异常值抑制
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2023.02.004
Qingqing Chen , Junfeng Yao

Background

Image matching is crucial in numerous computer vision tasks such as 3D reconstruction and simultaneous visual localization and mapping. The accuracy of the matching significantly impacted subsequent studies. Because of their local similarity, when image pairs contain comparable patterns but feature pairs are positioned differently, incorrect recognition can occur as global motion consistency is disregarded.

Methods

This study proposes an image-matching filtering algorithm based on global motion consistency. It can be used as a subsequent matching filter for the initial matching results generated by other matching algorithms based on the principle of motion smoothness. A particular matching algorithm can first be used to perform the initial matching; then, the rotation and movement information of the global feature vectors are combined to effectively identify outlier matches. The principle is that if the matching result is accurate, the feature vectors formed by any matched point should have similar rotation angles and moving distances. Thus, global motion direction and global motion distance consistencies were used to reject outliers caused by similar patterns in different locations.

Results

Four datasets were used to test the effectiveness of the proposed method. Three datasets with similar patterns in different locations were used to test the results for similar images that could easily be incorrectly matched by other algorithms, and one commonly used dataset was used to test the results for the general image-matching problem. The experimental results suggest that the proposed method is more accurate than other state-of-the-art algorithms in identifying mismatches in the initial matching set.

Conclusions

The proposed outlier rejection matching method can significantly improve the matching accuracy for similar images with locally similar feature pairs in different locations and can provide more accurate matching results for subsequent computer vision tasks.

背景图像匹配在许多计算机视觉任务中至关重要,例如三维重建和同步视觉定位和映射。匹配的准确性显著影响了后续的研究。由于它们的局部相似性,当图像对包含可比较的模式但特征对的定位不同时,由于忽略了全局运动一致性,可能会发生错误的识别。方法提出一种基于全局运动一致性的图像匹配滤波算法。它可以作为其他匹配算法基于运动平滑原理产生的初始匹配结果的后续匹配滤波器。可首先使用特定匹配算法来执行初始匹配;然后,结合全局特征向量的旋转和运动信息,有效识别离群匹配;其原理是,如果匹配结果准确,则任何匹配点所形成的特征向量应具有相似的旋转角度和移动距离。因此,使用全局运动方向和全局运动距离一致性来拒绝不同位置相似模式造成的异常值。结果利用4个数据集验证了该方法的有效性。使用三个在不同位置具有相似模式的数据集来测试容易被其他算法错误匹配的相似图像的结果,并使用一个常用数据集来测试一般图像匹配问题的结果。实验结果表明,该方法在识别初始匹配集中的不匹配方面比其他先进算法更准确。结论本文提出的离群值抑制匹配方法可以显著提高不同位置具有局部相似特征对的相似图像的匹配精度,为后续计算机视觉任务提供更准确的匹配结果。
{"title":"Outliers rejection in similar image matching","authors":"Qingqing Chen ,&nbsp;Junfeng Yao","doi":"10.1016/j.vrih.2023.02.004","DOIUrl":"https://doi.org/10.1016/j.vrih.2023.02.004","url":null,"abstract":"<div><h3>Background</h3><p>Image matching is crucial in numerous computer vision tasks such as 3D reconstruction and simultaneous visual localization and mapping. The accuracy of the matching significantly impacted subsequent studies. Because of their local similarity, when image pairs contain comparable patterns but feature pairs are positioned differently, incorrect recognition can occur as global motion consistency is disregarded.</p></div><div><h3>Methods</h3><p>This study proposes an image-matching filtering algorithm based on global motion consistency. It can be used as a subsequent matching filter for the initial matching results generated by other matching algorithms based on the principle of motion smoothness. A particular matching algorithm can first be used to perform the initial matching; then, the rotation and movement information of the global feature vectors are combined to effectively identify outlier matches. The principle is that if the matching result is accurate, the feature vectors formed by any matched point should have similar rotation angles and moving distances. Thus, global motion direction and global motion distance consistencies were used to reject outliers caused by similar patterns in different locations.</p></div><div><h3>Results</h3><p>Four datasets were used to test the effectiveness of the proposed method. Three datasets with similar patterns in different locations were used to test the results for similar images that could easily be incorrectly matched by other algorithms, and one commonly used dataset was used to test the results for the general image-matching problem. The experimental results suggest that the proposed method is more accurate than other state-of-the-art algorithms in identifying mismatches in the initial matching set.</p></div><div><h3>Conclusions</h3><p>The proposed outlier rejection matching method can significantly improve the matching accuracy for similar images with locally similar feature pairs in different locations and can provide more accurate matching results for subsequent computer vision tasks.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 171-187"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49866057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive spatio-temporal attention neural network for crossdatabase micro-expression recognition 跨数据库微表情识别的自适应时空注意神经网络
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.03.006
Yuhan RAN

Background

The use of micro-expression recognition to recognize human emotions is one of the most critical challenges in human-computer interaction applications. In recent years, cross-database micro-expression recognition (CDMER) has emerged as a significant challenge in micro-expression recognition and analysis. Because the training and testing data in CDMER come from different micro-expression databases, CDMER is more challenging than conventional micro-expression recognition.

Methods

In this paper, an adaptive spatio-temporal attention neural network (ASTANN) using an attention mechanism is presented to address this challenge. To this end, the micro-expression databases SMIC and CASME II are first preprocessed using an optical flow approach, which extracts motion information among video frames that represent discriminative features of micro-expression. After preprocessing, a novel adaptive framework with a spatiotemporal attention module was designed to assign spatial and temporal weights to enhance the most discriminative features. The deep neural network then extracts the cross-domain feature, in which the second-order statistics of the sample features in the source domain are aligned with those in the target domain by minimizing the correlation alignment (CORAL) loss such that the source and target databases share similar distributions.

Results

To evaluate the performance of ASTANN, experiments were conducted based on the SMIC and CASME II databases under the standard experimental evaluation protocol of CDMER. The experimental results demonstrate that ASTANN outperformed other methods in relevant crossdatabase tasks.

Conclusions

Extensive experiments were conducted on benchmark tasks, and the results show that ASTANN has superior performance compared with other approaches. This demonstrates the superiority of our method in solving the CDMER problem.

使用微表情识别识别人类情绪是人机交互应用中最关键的挑战之一。近年来,跨数据库微表情识别(CDMER)已成为微表情识别和分析领域的一个重大挑战。由于CDMER的训练和测试数据来自不同的微表情数据库,因此CDMER比传统的微表情识别更具挑战性。方法提出了一种基于注意机制的自适应时空注意神经网络(ASTANN)来解决这一问题。为此,首先利用光流方法对微表情数据库SMIC和CASME II进行预处理,提取代表微表情特征的视频帧之间的运动信息。经过预处理,设计了一种具有时空注意模块的自适应框架来分配空间和时间权重,以增强最具区别性的特征。然后,深度神经网络提取跨域特征,其中源域样本特征的二阶统计量与目标域特征的二阶统计量通过最小化相关对齐(CORAL)损失进行对齐,从而使源数据库和目标数据库共享相似的分布。结果为评价ASTANN的性能,在CDMER标准实验评价方案下,基于SMIC和CASME II数据库进行了实验。实验结果表明,ASTANN在相关的跨数据库任务中优于其他方法。结论在基准任务上进行了大量的实验,结果表明,与其他方法相比,ASTANN具有优越的性能。这证明了我们的方法在解决CDMER问题上的优越性。
{"title":"Adaptive spatio-temporal attention neural network for crossdatabase micro-expression recognition","authors":"Yuhan RAN","doi":"10.1016/j.vrih.2022.03.006","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.03.006","url":null,"abstract":"<div><h3>Background</h3><p>The use of micro-expression recognition to recognize human emotions is one of the most critical challenges in human-computer interaction applications. In recent years, cross-database micro-expression recognition (CDMER) has emerged as a significant challenge in micro-expression recognition and analysis. Because the training and testing data in CDMER come from different micro-expression databases, CDMER is more challenging than conventional micro-expression recognition.</p></div><div><h3>Methods</h3><p>In this paper, an adaptive spatio-temporal attention neural network (ASTANN) using an attention mechanism is presented to address this challenge. To this end, the micro-expression databases SMIC and CASME II are first preprocessed using an optical flow approach, which extracts motion information among video frames that represent discriminative features of micro-expression. After preprocessing, a novel adaptive framework with a spatiotemporal attention module was designed to assign spatial and temporal weights to enhance the most discriminative features. The deep neural network then extracts the cross-domain feature, in which the second-order statistics of the sample features in the source domain are aligned with those in the target domain by minimizing the correlation alignment (CORAL) loss such that the source and target databases share similar distributions.</p></div><div><h3>Results</h3><p>To evaluate the performance of ASTANN, experiments were conducted based on the SMIC and CASME II databases under the standard experimental evaluation protocol of CDMER. The experimental results demonstrate that ASTANN outperformed other methods in relevant crossdatabase tasks.</p></div><div><h3>Conclusions</h3><p>Extensive experiments were conducted on benchmark tasks, and the results show that ASTANN has superior performance compared with other approaches. This demonstrates the superiority of our method in solving the CDMER problem.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 142-156"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49891619","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Web-based Mixed Reality Video Fusion with Remote Rendering 基于web的混合现实视频融合与远程渲染
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.03.005
Qiang Zhou, Zhong Zhou

Mixed Reality (MR) video fusion system fuses video imagery with 3D scenes. It makes the scene much more realistic and helps the users understand the video contents and temporalspatial correlation between them, thus reducing the user’s cognitive load. Nowadays, MR video fusion has been used in various applications. However, video fusion systems require powerful client machines because video streaming delivery, stitching, and rendering are computation-intensive. Moreover, huge bandwidth usage is also another critical factor that affects the scalability of video fusion systems. The framework proposed in this paper overcomes this client limitation by utilizing remote rendering. Furthermore, the framework we built is based on browsers. Therefore, the user could try the MR video fusion system with a laptop or even pad, no extra plug-ins or application programs need to be installed. Several experiments on diverse metrics demonstrate the effectiveness of the proposed framework.

混合现实(MR)视频融合系统将视频图像与3D场景融合在一起。它使场景更加逼真,帮助用户理解视频内容和它们之间的时空相关性,从而减少用户的认知负荷。目前,磁共振视频融合已被广泛应用。然而,视频融合系统需要强大的客户端机器,因为视频流传输、拼接和渲染是计算密集型的。此外,巨大的带宽占用也是影响视频融合系统可扩展性的另一个关键因素。本文提出的框架利用远程渲染技术克服了这种客户端限制。此外,我们构建的框架是基于浏览器的。因此,用户可以在笔记本电脑甚至pad上试用MR视频融合系统,无需额外安装插件或应用程序。在不同指标上的几个实验证明了所提出框架的有效性。
{"title":"Web-based Mixed Reality Video Fusion with Remote Rendering","authors":"Qiang Zhou,&nbsp;Zhong Zhou","doi":"10.1016/j.vrih.2022.03.005","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.03.005","url":null,"abstract":"<div><p>Mixed Reality (MR) video fusion system fuses video imagery with 3D scenes. It makes the scene much more realistic and helps the users understand the video contents and temporalspatial correlation between them, thus reducing the user’s cognitive load. Nowadays, MR video fusion has been used in various applications. However, video fusion systems require powerful client machines because video streaming delivery, stitching, and rendering are computation-intensive. Moreover, huge bandwidth usage is also another critical factor that affects the scalability of video fusion systems. The framework proposed in this paper overcomes this client limitation by utilizing remote rendering. Furthermore, the framework we built is based on browsers. Therefore, the user could try the MR video fusion system with a laptop or even pad, no extra plug-ins or application programs need to be installed. Several experiments on diverse metrics demonstrate the effectiveness of the proposed framework.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 188-199"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49891615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Compression of Surface Texture Acceleration Signal Based on Spectrum Characteristics 基于频谱特征的表面纹理加速信号压缩
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.01.006
Dongyan Nie , Xiaoying Sun

Background

Adequate-data collection could enhance the realism of surface texture haptic online-rendering or offline-playback. A parallel challenge is how to reduce communication delays and improve storage space utilization.

Methods

Based on the similarity of the short-term amplitude spectrumtrend, this paper proposes a frequency-domain compression method. A compression framework is designed, firstly to map the amplitude spectrum into a trend similarity grayscale image, compress it with the stillpicture-compression method, and then to adaptively encode the maximum amplitude and part of the initial phase of each time-window, achieving the final compression.

Results

The comparison between the original signal and the recovered signal shows that when the time-frequency similarity is 90%, the average compression ratio of our method is 9.85% in the case of a single interact point. The subjective score for the similarity reached an excellent level, with an average score of 87.85.

Conclusions

Our method can be used for offline compression of vibrotactile data. For the case of multi-interact points in space, the trend similarity grayscale image can be reused, and the compression ratio is further reduced.

背景充分的数据采集可以提高表面纹理触觉在线渲染或离线回放的真实感。一个并行的挑战是如何减少通信延迟和提高存储空间利用率。方法基于短期振幅谱趋势的相似性,提出了一种频域压缩方法。设计了压缩框架,首先将振幅谱映射成趋势相似灰度图像,采用静态图像压缩方法进行压缩,然后对每个时间窗的最大振幅和部分初始相位进行自适应编码,实现最终压缩。结果原始信号与恢复信号的对比表明,当时频相似度为90%时,在单交互点情况下,本文方法的平均压缩比为9.85%。相似度主观得分达到优秀水平,平均得分87.85。结论sour方法可用于振动触觉数据的离线压缩。对于空间中存在多交互点的情况,趋势相似度灰度图像可以重复使用,压缩比进一步降低。
{"title":"Compression of Surface Texture Acceleration Signal Based on Spectrum Characteristics","authors":"Dongyan Nie ,&nbsp;Xiaoying Sun","doi":"10.1016/j.vrih.2022.01.006","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.01.006","url":null,"abstract":"<div><h3>Background</h3><p>Adequate-data collection could enhance the realism of surface texture haptic online-rendering or offline-playback. A parallel challenge is how to reduce communication delays and improve storage space utilization.</p></div><div><h3>Methods</h3><p>Based on the similarity of the short-term amplitude spectrumtrend, this paper proposes a frequency-domain compression method. A compression framework is designed, firstly to map the amplitude spectrum into a trend similarity grayscale image, compress it with the stillpicture-compression method, and then to adaptively encode the maximum amplitude and part of the initial phase of each time-window, achieving the final compression.</p></div><div><h3>Results</h3><p>The comparison between the original signal and the recovered signal shows that when the time-frequency similarity is 90%, the average compression ratio of our method is 9.85% in the case of a single interact point. The subjective score for the similarity reached an excellent level, with an average score of 87.85.</p></div><div><h3>Conclusions</h3><p>Our method can be used for offline compression of vibrotactile data. For the case of multi-interact points in space, the trend similarity grayscale image can be reused, and the compression ratio is further reduced.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 110-123"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49891620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MSSTNet: Multi-scale facial videos pulse extraction network based on separable spatiotemporal convolution and dimension separable attention 基于可分时空卷积和维数可分注意力的多尺度面部视频脉冲提取网络
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.07.001
Changchen Zhao , Hongsheng Wang , Yuanjing Feng

Background

Using remote photoplethysmography (rPPG) to estimate blood volume pulse in a non-contact way is an active research topic in recent years. Existing methods are mainly based on the single-scale region of interest (ROI). However, some noise signals that are not easily separated in single-scale space can be easily separated in multi-scale space. In addition, existing spatiotemporal networks mainly focus on local spatiotemporal information and lack emphasis on temporal information which is crucial in pulse extraction problems, resulting in insufficient spatiotemporal feature modeling.

Methods

This paper proposes a multi-scale facial video pulse extraction network based on separable spatiotemporal convolution and dimension separable attention. First, in order to solve the problem of single-scale ROI, we construct a multi-scale feature space for initial signal separation. Secondly, separable spatiotemporal convolution and dimension separable attention are designed for efficient spatiotemporal correlation modeling, which increases the information interaction between long-span time and space dimensions and puts more emphasis on temporal features.

Results

The signal-to-noise ratio (SNR) of the proposed network reaches 9.58 dB on the PURE dataset and 6.77 dB on the UBFC-rPPG dataset, which outperforms state-of-the-art algorithms.

Conclusions

Results show that fusing multi-scale signals generally obtains better results than methods based on the only single-scale signal. The proposed separable spatiotemporal convolution and dimension separable attention mechanism contributes to more accurate pulse signal extraction.

利用远程光容积脉搏波(rPPG)非接触式测量血容量脉搏是近年来研究的热点。现有方法主要基于单尺度感兴趣区域(ROI)。然而,一些在单尺度空间中不易分离的噪声信号在多尺度空间中却很容易分离。此外,现有的时空网络主要关注局部时空信息,缺乏对脉冲提取问题中至关重要的时间信息的重视,导致时空特征建模不足。方法提出了一种基于可分时空卷积和维数可分注意力的多尺度面部视频脉冲提取网络。首先,为了解决单尺度ROI问题,构建多尺度特征空间进行初始信号分离;其次,设计了可分时空卷积和可分维度注意的高效时空关联建模方法,增加了大跨度时空维度之间的信息交互,更加强调时间特征;结果该网络在PURE数据集上的信噪比达到9.58 dB,在UBFC-rPPG数据集上的信噪比达到6.77 dB,优于现有算法。结论多尺度信号融合总体上优于单尺度信号融合。提出的可分时空卷积和可分维注意机制有助于提高脉冲信号的提取精度。
{"title":"MSSTNet: Multi-scale facial videos pulse extraction network based on separable spatiotemporal convolution and dimension separable attention","authors":"Changchen Zhao ,&nbsp;Hongsheng Wang ,&nbsp;Yuanjing Feng","doi":"10.1016/j.vrih.2022.07.001","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.07.001","url":null,"abstract":"<div><h3>Background</h3><p>Using remote photoplethysmography (rPPG) to estimate blood volume pulse in a non-contact way is an active research topic in recent years. Existing methods are mainly based on the single-scale region of interest (ROI). However, some noise signals that are not easily separated in single-scale space can be easily separated in multi-scale space. In addition, existing spatiotemporal networks mainly focus on local spatiotemporal information and lack emphasis on temporal information which is crucial in pulse extraction problems, resulting in insufficient spatiotemporal feature modeling.</p></div><div><h3>Methods</h3><p>This paper proposes a multi-scale facial video pulse extraction network based on separable spatiotemporal convolution and dimension separable attention. First, in order to solve the problem of single-scale ROI, we construct a multi-scale feature space for initial signal separation. Secondly, separable spatiotemporal convolution and dimension separable attention are designed for efficient spatiotemporal correlation modeling, which increases the information interaction between long-span time and space dimensions and puts more emphasis on temporal features.</p></div><div><h3>Results</h3><p>The signal-to-noise ratio (SNR) of the proposed network reaches 9.58 dB on the PURE dataset and 6.77 dB on the UBFC-rPPG dataset, which outperforms state-of-the-art algorithms.</p></div><div><h3>Conclusions</h3><p>Results show that fusing multi-scale signals generally obtains better results than methods based on the only single-scale signal. The proposed separable spatiotemporal convolution and dimension separable attention mechanism contributes to more accurate pulse signal extraction.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 124-141"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49891621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Intelligent Fire Information System Based on 3D GIS 基于三维GIS的智能火灾信息系统
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.07.002
Jinxing Hu , Zhihan Lv , Diping Yuan , Bing He , Dongmei Yan

This work aims to build a comprehensive and effective fire emergency management system based on the Internet of Things (IoT) and achieve an actual intelligent fire rescue. A smart fire protection information system was designed based on the IoT. A detailed analysis was conducted on the problem of rescue vehicle scheduling and the evacuation of trapped persons in the process of fire rescue. The intelligent fire visualization platform based on the three-dimensional (3D) Geographic Information Science (GIS) covers project overview, equipment status, equipment classification, equipment alarm information, alarm classification, alarm statistics, equipment account information, and other modules. The live video accessed through the visual interface can clearly identify the stage of the fire, which facilitates the arrangement of rescue equipment and personnel. The vehicle scheduling model in the system primarily used two objective functions to solve the Pareto Non-Dominated Solution Set Optimization: emergency rescue time and the number of vehicles. In addition, an evacuation path optimization method based on the Improved Ant Colony (IAC) algorithm was designed to realize the dynamic optimization of building fire evacuation paths. The experimental results indicate that all the values of detection signals were significantly larger in the smoldering fire scene at t = 17s than the initial value. In addition, the probability of smoldering fire and the probability of open fire were relatively large according to the probability function of the corresponding fire situation, demonstrating that this model could detect fire. The IAC algorithm reported here avoided the passages near the fire and spreading areas as much as possible and took the safety of the trapped persons as the premise when planning the evacuation route. Therefore, the IoT-based fire information system has important value for ensuring fire safety and carrying out emergency rescue and is worthy of popularization and application.

本工作旨在构建基于物联网的全面有效的消防应急管理体系,实现真正的智能消防救援。设计了基于物联网的智能消防信息系统。对火灾救援过程中救援车辆调度和被困人员疏散问题进行了详细分析。基于三维地理信息科学(GIS)的智能消防可视化平台,包括工程概况、设备状态、设备分类、设备报警信息、报警分类、报警统计、设备台账信息等模块。通过可视化界面访问的现场视频可以清晰地识别火灾的阶段,方便安排救援设备和人员。系统中的车辆调度模型主要使用两个目标函数来求解Pareto非支配解集优化问题:紧急救援时间和车辆数量。此外,设计了一种基于改进蚁群(IAC)算法的疏散路径优化方法,实现了建筑火灾疏散路径的动态优化。实验结果表明,在t = 17s时,阴燃现场的所有探测信号值都明显大于初始值。此外,根据对应火灾情况的概率函数,阴燃发生的概率和明火发生的概率都比较大,说明该模型能够探测到火灾。本文报道的IAC算法在规划疏散路线时,尽可能避开靠近火场和蔓延区域的通道,以被困人员的安全为前提。因此,基于物联网的消防信息系统对于保障消防安全和开展应急救援具有重要价值,值得推广应用。
{"title":"Intelligent Fire Information System Based on 3D GIS","authors":"Jinxing Hu ,&nbsp;Zhihan Lv ,&nbsp;Diping Yuan ,&nbsp;Bing He ,&nbsp;Dongmei Yan","doi":"10.1016/j.vrih.2022.07.002","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.07.002","url":null,"abstract":"<div><p>This work aims to build a comprehensive and effective fire emergency management system based on the Internet of Things (IoT) and achieve an actual intelligent fire rescue. A smart fire protection information system was designed based on the IoT. A detailed analysis was conducted on the problem of rescue vehicle scheduling and the evacuation of trapped persons in the process of fire rescue. The intelligent fire visualization platform based on the three-dimensional (3D) Geographic Information Science (GIS) covers project overview, equipment status, equipment classification, equipment alarm information, alarm classification, alarm statistics, equipment account information, and other modules. The live video accessed through the visual interface can clearly identify the stage of the fire, which facilitates the arrangement of rescue equipment and personnel. The vehicle scheduling model in the system primarily used two objective functions to solve the Pareto Non-Dominated Solution Set Optimization: emergency rescue time and the number of vehicles. In addition, an evacuation path optimization method based on the Improved Ant Colony (IAC) algorithm was designed to realize the dynamic optimization of building fire evacuation paths. The experimental results indicate that all the values of detection signals were significantly larger in the smoldering fire scene at t = 17s than the initial value. In addition, the probability of smoldering fire and the probability of open fire were relatively large according to the probability function of the corresponding fire situation, demonstrating that this model could detect fire. The IAC algorithm reported here avoided the passages near the fire and spreading areas as much as possible and took the safety of the trapped persons as the premise when planning the evacuation route. Therefore, the IoT-based fire information system has important value for ensuring fire safety and carrying out emergency rescue and is worthy of popularization and application.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 93-109"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49866020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hardware—A New Open Access Journal 硬件——一个新的开放获取期刊
Q1 Computer Science Pub Date : 2023-03-30 DOI: 10.3390/hardware1010001
Peter C. Hauser
Hardware (ISSN 2813-6640) [...]
{"title":"Hardware—A New Open Access Journal","authors":"Peter C. Hauser","doi":"10.3390/hardware1010001","DOIUrl":"https://doi.org/10.3390/hardware1010001","url":null,"abstract":"Hardware (ISSN 2813-6640) [...]","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"191 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85126224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Virtual Reality Intelligent Hardware
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1