首页 > 最新文献

Virtual Reality Intelligent Hardware最新文献

英文 中文
Web-based Mixed Reality Video Fusion with Remote Rendering 基于web的混合现实视频融合与远程渲染
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.03.005
Qiang Zhou, Zhong Zhou

Mixed Reality (MR) video fusion system fuses video imagery with 3D scenes. It makes the scene much more realistic and helps the users understand the video contents and temporalspatial correlation between them, thus reducing the user’s cognitive load. Nowadays, MR video fusion has been used in various applications. However, video fusion systems require powerful client machines because video streaming delivery, stitching, and rendering are computation-intensive. Moreover, huge bandwidth usage is also another critical factor that affects the scalability of video fusion systems. The framework proposed in this paper overcomes this client limitation by utilizing remote rendering. Furthermore, the framework we built is based on browsers. Therefore, the user could try the MR video fusion system with a laptop or even pad, no extra plug-ins or application programs need to be installed. Several experiments on diverse metrics demonstrate the effectiveness of the proposed framework.

混合现实(MR)视频融合系统将视频图像与3D场景融合在一起。它使场景更加逼真,帮助用户理解视频内容和它们之间的时空相关性,从而减少用户的认知负荷。目前,磁共振视频融合已被广泛应用。然而,视频融合系统需要强大的客户端机器,因为视频流传输、拼接和渲染是计算密集型的。此外,巨大的带宽占用也是影响视频融合系统可扩展性的另一个关键因素。本文提出的框架利用远程渲染技术克服了这种客户端限制。此外,我们构建的框架是基于浏览器的。因此,用户可以在笔记本电脑甚至pad上试用MR视频融合系统,无需额外安装插件或应用程序。在不同指标上的几个实验证明了所提出框架的有效性。
{"title":"Web-based Mixed Reality Video Fusion with Remote Rendering","authors":"Qiang Zhou,&nbsp;Zhong Zhou","doi":"10.1016/j.vrih.2022.03.005","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.03.005","url":null,"abstract":"<div><p>Mixed Reality (MR) video fusion system fuses video imagery with 3D scenes. It makes the scene much more realistic and helps the users understand the video contents and temporalspatial correlation between them, thus reducing the user’s cognitive load. Nowadays, MR video fusion has been used in various applications. However, video fusion systems require powerful client machines because video streaming delivery, stitching, and rendering are computation-intensive. Moreover, huge bandwidth usage is also another critical factor that affects the scalability of video fusion systems. The framework proposed in this paper overcomes this client limitation by utilizing remote rendering. Furthermore, the framework we built is based on browsers. Therefore, the user could try the MR video fusion system with a laptop or even pad, no extra plug-ins or application programs need to be installed. Several experiments on diverse metrics demonstrate the effectiveness of the proposed framework.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 188-199"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49891615","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Compression of Surface Texture Acceleration Signal Based on Spectrum Characteristics 基于频谱特征的表面纹理加速信号压缩
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.01.006
Dongyan Nie , Xiaoying Sun

Background

Adequate-data collection could enhance the realism of surface texture haptic online-rendering or offline-playback. A parallel challenge is how to reduce communication delays and improve storage space utilization.

Methods

Based on the similarity of the short-term amplitude spectrumtrend, this paper proposes a frequency-domain compression method. A compression framework is designed, firstly to map the amplitude spectrum into a trend similarity grayscale image, compress it with the stillpicture-compression method, and then to adaptively encode the maximum amplitude and part of the initial phase of each time-window, achieving the final compression.

Results

The comparison between the original signal and the recovered signal shows that when the time-frequency similarity is 90%, the average compression ratio of our method is 9.85% in the case of a single interact point. The subjective score for the similarity reached an excellent level, with an average score of 87.85.

Conclusions

Our method can be used for offline compression of vibrotactile data. For the case of multi-interact points in space, the trend similarity grayscale image can be reused, and the compression ratio is further reduced.

背景充分的数据采集可以提高表面纹理触觉在线渲染或离线回放的真实感。一个并行的挑战是如何减少通信延迟和提高存储空间利用率。方法基于短期振幅谱趋势的相似性,提出了一种频域压缩方法。设计了压缩框架,首先将振幅谱映射成趋势相似灰度图像,采用静态图像压缩方法进行压缩,然后对每个时间窗的最大振幅和部分初始相位进行自适应编码,实现最终压缩。结果原始信号与恢复信号的对比表明,当时频相似度为90%时,在单交互点情况下,本文方法的平均压缩比为9.85%。相似度主观得分达到优秀水平,平均得分87.85。结论sour方法可用于振动触觉数据的离线压缩。对于空间中存在多交互点的情况,趋势相似度灰度图像可以重复使用,压缩比进一步降低。
{"title":"Compression of Surface Texture Acceleration Signal Based on Spectrum Characteristics","authors":"Dongyan Nie ,&nbsp;Xiaoying Sun","doi":"10.1016/j.vrih.2022.01.006","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.01.006","url":null,"abstract":"<div><h3>Background</h3><p>Adequate-data collection could enhance the realism of surface texture haptic online-rendering or offline-playback. A parallel challenge is how to reduce communication delays and improve storage space utilization.</p></div><div><h3>Methods</h3><p>Based on the similarity of the short-term amplitude spectrumtrend, this paper proposes a frequency-domain compression method. A compression framework is designed, firstly to map the amplitude spectrum into a trend similarity grayscale image, compress it with the stillpicture-compression method, and then to adaptively encode the maximum amplitude and part of the initial phase of each time-window, achieving the final compression.</p></div><div><h3>Results</h3><p>The comparison between the original signal and the recovered signal shows that when the time-frequency similarity is 90%, the average compression ratio of our method is 9.85% in the case of a single interact point. The subjective score for the similarity reached an excellent level, with an average score of 87.85.</p></div><div><h3>Conclusions</h3><p>Our method can be used for offline compression of vibrotactile data. For the case of multi-interact points in space, the trend similarity grayscale image can be reused, and the compression ratio is further reduced.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 110-123"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49891620","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MSSTNet: Multi-scale facial videos pulse extraction network based on separable spatiotemporal convolution and dimension separable attention 基于可分时空卷积和维数可分注意力的多尺度面部视频脉冲提取网络
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.07.001
Changchen Zhao , Hongsheng Wang , Yuanjing Feng

Background

Using remote photoplethysmography (rPPG) to estimate blood volume pulse in a non-contact way is an active research topic in recent years. Existing methods are mainly based on the single-scale region of interest (ROI). However, some noise signals that are not easily separated in single-scale space can be easily separated in multi-scale space. In addition, existing spatiotemporal networks mainly focus on local spatiotemporal information and lack emphasis on temporal information which is crucial in pulse extraction problems, resulting in insufficient spatiotemporal feature modeling.

Methods

This paper proposes a multi-scale facial video pulse extraction network based on separable spatiotemporal convolution and dimension separable attention. First, in order to solve the problem of single-scale ROI, we construct a multi-scale feature space for initial signal separation. Secondly, separable spatiotemporal convolution and dimension separable attention are designed for efficient spatiotemporal correlation modeling, which increases the information interaction between long-span time and space dimensions and puts more emphasis on temporal features.

Results

The signal-to-noise ratio (SNR) of the proposed network reaches 9.58 dB on the PURE dataset and 6.77 dB on the UBFC-rPPG dataset, which outperforms state-of-the-art algorithms.

Conclusions

Results show that fusing multi-scale signals generally obtains better results than methods based on the only single-scale signal. The proposed separable spatiotemporal convolution and dimension separable attention mechanism contributes to more accurate pulse signal extraction.

利用远程光容积脉搏波(rPPG)非接触式测量血容量脉搏是近年来研究的热点。现有方法主要基于单尺度感兴趣区域(ROI)。然而,一些在单尺度空间中不易分离的噪声信号在多尺度空间中却很容易分离。此外,现有的时空网络主要关注局部时空信息,缺乏对脉冲提取问题中至关重要的时间信息的重视,导致时空特征建模不足。方法提出了一种基于可分时空卷积和维数可分注意力的多尺度面部视频脉冲提取网络。首先,为了解决单尺度ROI问题,构建多尺度特征空间进行初始信号分离;其次,设计了可分时空卷积和可分维度注意的高效时空关联建模方法,增加了大跨度时空维度之间的信息交互,更加强调时间特征;结果该网络在PURE数据集上的信噪比达到9.58 dB,在UBFC-rPPG数据集上的信噪比达到6.77 dB,优于现有算法。结论多尺度信号融合总体上优于单尺度信号融合。提出的可分时空卷积和可分维注意机制有助于提高脉冲信号的提取精度。
{"title":"MSSTNet: Multi-scale facial videos pulse extraction network based on separable spatiotemporal convolution and dimension separable attention","authors":"Changchen Zhao ,&nbsp;Hongsheng Wang ,&nbsp;Yuanjing Feng","doi":"10.1016/j.vrih.2022.07.001","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.07.001","url":null,"abstract":"<div><h3>Background</h3><p>Using remote photoplethysmography (rPPG) to estimate blood volume pulse in a non-contact way is an active research topic in recent years. Existing methods are mainly based on the single-scale region of interest (ROI). However, some noise signals that are not easily separated in single-scale space can be easily separated in multi-scale space. In addition, existing spatiotemporal networks mainly focus on local spatiotemporal information and lack emphasis on temporal information which is crucial in pulse extraction problems, resulting in insufficient spatiotemporal feature modeling.</p></div><div><h3>Methods</h3><p>This paper proposes a multi-scale facial video pulse extraction network based on separable spatiotemporal convolution and dimension separable attention. First, in order to solve the problem of single-scale ROI, we construct a multi-scale feature space for initial signal separation. Secondly, separable spatiotemporal convolution and dimension separable attention are designed for efficient spatiotemporal correlation modeling, which increases the information interaction between long-span time and space dimensions and puts more emphasis on temporal features.</p></div><div><h3>Results</h3><p>The signal-to-noise ratio (SNR) of the proposed network reaches 9.58 dB on the PURE dataset and 6.77 dB on the UBFC-rPPG dataset, which outperforms state-of-the-art algorithms.</p></div><div><h3>Conclusions</h3><p>Results show that fusing multi-scale signals generally obtains better results than methods based on the only single-scale signal. The proposed separable spatiotemporal convolution and dimension separable attention mechanism contributes to more accurate pulse signal extraction.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 124-141"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49891621","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Intelligent Fire Information System Based on 3D GIS 基于三维GIS的智能火灾信息系统
Q1 Computer Science Pub Date : 2023-04-01 DOI: 10.1016/j.vrih.2022.07.002
Jinxing Hu , Zhihan Lv , Diping Yuan , Bing He , Dongmei Yan

This work aims to build a comprehensive and effective fire emergency management system based on the Internet of Things (IoT) and achieve an actual intelligent fire rescue. A smart fire protection information system was designed based on the IoT. A detailed analysis was conducted on the problem of rescue vehicle scheduling and the evacuation of trapped persons in the process of fire rescue. The intelligent fire visualization platform based on the three-dimensional (3D) Geographic Information Science (GIS) covers project overview, equipment status, equipment classification, equipment alarm information, alarm classification, alarm statistics, equipment account information, and other modules. The live video accessed through the visual interface can clearly identify the stage of the fire, which facilitates the arrangement of rescue equipment and personnel. The vehicle scheduling model in the system primarily used two objective functions to solve the Pareto Non-Dominated Solution Set Optimization: emergency rescue time and the number of vehicles. In addition, an evacuation path optimization method based on the Improved Ant Colony (IAC) algorithm was designed to realize the dynamic optimization of building fire evacuation paths. The experimental results indicate that all the values of detection signals were significantly larger in the smoldering fire scene at t = 17s than the initial value. In addition, the probability of smoldering fire and the probability of open fire were relatively large according to the probability function of the corresponding fire situation, demonstrating that this model could detect fire. The IAC algorithm reported here avoided the passages near the fire and spreading areas as much as possible and took the safety of the trapped persons as the premise when planning the evacuation route. Therefore, the IoT-based fire information system has important value for ensuring fire safety and carrying out emergency rescue and is worthy of popularization and application.

本工作旨在构建基于物联网的全面有效的消防应急管理体系,实现真正的智能消防救援。设计了基于物联网的智能消防信息系统。对火灾救援过程中救援车辆调度和被困人员疏散问题进行了详细分析。基于三维地理信息科学(GIS)的智能消防可视化平台,包括工程概况、设备状态、设备分类、设备报警信息、报警分类、报警统计、设备台账信息等模块。通过可视化界面访问的现场视频可以清晰地识别火灾的阶段,方便安排救援设备和人员。系统中的车辆调度模型主要使用两个目标函数来求解Pareto非支配解集优化问题:紧急救援时间和车辆数量。此外,设计了一种基于改进蚁群(IAC)算法的疏散路径优化方法,实现了建筑火灾疏散路径的动态优化。实验结果表明,在t = 17s时,阴燃现场的所有探测信号值都明显大于初始值。此外,根据对应火灾情况的概率函数,阴燃发生的概率和明火发生的概率都比较大,说明该模型能够探测到火灾。本文报道的IAC算法在规划疏散路线时,尽可能避开靠近火场和蔓延区域的通道,以被困人员的安全为前提。因此,基于物联网的消防信息系统对于保障消防安全和开展应急救援具有重要价值,值得推广应用。
{"title":"Intelligent Fire Information System Based on 3D GIS","authors":"Jinxing Hu ,&nbsp;Zhihan Lv ,&nbsp;Diping Yuan ,&nbsp;Bing He ,&nbsp;Dongmei Yan","doi":"10.1016/j.vrih.2022.07.002","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.07.002","url":null,"abstract":"<div><p>This work aims to build a comprehensive and effective fire emergency management system based on the Internet of Things (IoT) and achieve an actual intelligent fire rescue. A smart fire protection information system was designed based on the IoT. A detailed analysis was conducted on the problem of rescue vehicle scheduling and the evacuation of trapped persons in the process of fire rescue. The intelligent fire visualization platform based on the three-dimensional (3D) Geographic Information Science (GIS) covers project overview, equipment status, equipment classification, equipment alarm information, alarm classification, alarm statistics, equipment account information, and other modules. The live video accessed through the visual interface can clearly identify the stage of the fire, which facilitates the arrangement of rescue equipment and personnel. The vehicle scheduling model in the system primarily used two objective functions to solve the Pareto Non-Dominated Solution Set Optimization: emergency rescue time and the number of vehicles. In addition, an evacuation path optimization method based on the Improved Ant Colony (IAC) algorithm was designed to realize the dynamic optimization of building fire evacuation paths. The experimental results indicate that all the values of detection signals were significantly larger in the smoldering fire scene at t = 17s than the initial value. In addition, the probability of smoldering fire and the probability of open fire were relatively large according to the probability function of the corresponding fire situation, demonstrating that this model could detect fire. The IAC algorithm reported here avoided the passages near the fire and spreading areas as much as possible and took the safety of the trapped persons as the premise when planning the evacuation route. Therefore, the IoT-based fire information system has important value for ensuring fire safety and carrying out emergency rescue and is worthy of popularization and application.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 2","pages":"Pages 93-109"},"PeriodicalIF":0.0,"publicationDate":"2023-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49866020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hardware—A New Open Access Journal 硬件——一个新的开放获取期刊
Q1 Computer Science Pub Date : 2023-03-30 DOI: 10.3390/hardware1010001
Peter C. Hauser
Hardware (ISSN 2813-6640) [...]
{"title":"Hardware—A New Open Access Journal","authors":"Peter C. Hauser","doi":"10.3390/hardware1010001","DOIUrl":"https://doi.org/10.3390/hardware1010001","url":null,"abstract":"Hardware (ISSN 2813-6640) [...]","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"191 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85126224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Publisher’s Note: Hardware—A New Open Access Journal 出版商注:硬件——一种新的开放获取期刊
Q1 Computer Science Pub Date : 2023-03-30 DOI: 10.3390/hardware1010002
Liliane Auwerter
The development of new hardware has never been as accessible as it is today [...]
新硬件的开发从来没有像今天这样容易获得。
{"title":"Publisher’s Note: Hardware—A New Open Access Journal","authors":"Liliane Auwerter","doi":"10.3390/hardware1010002","DOIUrl":"https://doi.org/10.3390/hardware1010002","url":null,"abstract":"The development of new hardware has never been as accessible as it is today [...]","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"38 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2023-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74840471","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Transformer Architecture based mutual attention for Image Anomaly Detection 一种基于互感器结构的图像异常检测相互注意
Q1 Computer Science Pub Date : 2023-02-01 DOI: 10.1016/j.vrih.2022.07.006
Mengting Zhang, Xiuxia Tian

Background

Image anomaly detection is a popular task in computer graphics, which is widely used in industrial fields. Previous works that address this problem often train CNN-based (e.g. Auto-Encoder, GANs) models to reconstruct covered parts of input images and calculate the difference between the input and the reconstructed image. However, convolutional operations are good at extracting local features making it difficult to identify larger image anomalies. To this end, we propose a transformer architecture based on mutual attention for image anomaly separation. This architecture can capture long-term dependencies and fuse local features with global features to facilitate better image anomaly detection. Our method was extensively evaluated on several benchmarks, and experimental results showed that it improved detection capability by 3.1% and localization capability by 1.0% compared with state-of-the-art reconstruction-based methods.

背景图像异常检测是计算机图形学中的一项热门任务,在工业领域有着广泛的应用。解决这个问题的先前工作通常训练基于CNN的(例如,自动编码器,GANs)模型来重建输入图像的覆盖部分,并计算输入和重建图像之间的差。然而,卷积运算善于提取局部特征,这使得识别更大的图像异常变得困难。为此,我们提出了一种基于相互关注的图像异常分离转换器架构。该架构可以捕获长期相关性,并将局部特征与全局特征融合,以便于更好地检测图像异常。我们的方法在几个基准上进行了广泛的评估,实验结果表明,与最先进的基于重建的方法相比,它的检测能力提高了3.1%,定位能力提高了1.0%。
{"title":"A Transformer Architecture based mutual attention for Image Anomaly Detection","authors":"Mengting Zhang,&nbsp;Xiuxia Tian","doi":"10.1016/j.vrih.2022.07.006","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.07.006","url":null,"abstract":"<div><h3>Background</h3><p>Image anomaly detection is a popular task in computer graphics, which is widely used in industrial fields. Previous works that address this problem often train CNN-based (e.g. Auto-Encoder, GANs) models to reconstruct covered parts of input images and calculate the difference between the input and the reconstructed image. However, convolutional operations are good at extracting local features making it difficult to identify larger image anomalies. To this end, we propose a transformer architecture based on mutual attention for image anomaly separation. This architecture can capture long-term dependencies and fuse local features with global features to facilitate better image anomaly detection. Our method was extensively evaluated on several benchmarks, and experimental results showed that it improved detection capability by 3.1% and localization capability by 1.0% compared with state-of-the-art reconstruction-based methods.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 1","pages":"Pages 57-67"},"PeriodicalIF":0.0,"publicationDate":"2023-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49830409","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
View Interpolation Networks for Reproducing Material Appearance of Specular Objects 再现镜面反射物体材料外观的视图插值网络
Q1 Computer Science Pub Date : 2023-02-01 DOI: 10.1016/j.vrih.2022.11.001
Chihiro Hoshizawa, Takashi Komuro

In this study, we propose view interpolation networks to reproduce changes in the brightness of an object's surface depending on the viewing direction, which is important in reproducing the material appearance of a real object. We use an original and a modified version of U-Net for image transformation. The networks were trained to generate images from intermediate viewpoints of four cameras placed at the corners of a square. We conducted an experiment with three different combinations of methods and training data formats. We found that it is best to input the coordinates of the viewpoints together with the four camera images and to use images from random viewpoints as the training data.

在这项研究中,我们提出了视图插值网络来再现物体表面亮度随观看方向的变化,这对再现真实物体的材料外观很重要。我们使用U-Net的原始版本和修改版本进行图像转换。这些网络被训练成从放置在正方形角落的四个相机的中间视点生成图像。我们用三种不同的方法和训练数据格式组合进行了一项实验。我们发现,最好将视点的坐标与四个相机图像一起输入,并使用来自随机视点的图像作为训练数据。
{"title":"View Interpolation Networks for Reproducing Material Appearance of Specular Objects","authors":"Chihiro Hoshizawa,&nbsp;Takashi Komuro","doi":"10.1016/j.vrih.2022.11.001","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.11.001","url":null,"abstract":"<div><p>In this study, we propose view interpolation networks to reproduce changes in the brightness of an object's surface depending on the viewing direction, which is important in reproducing the material appearance of a real object. We use an original and a modified version of U-Net for image transformation. The networks were trained to generate images from intermediate viewpoints of four cameras placed at the corners of a square. We conducted an experiment with three different combinations of methods and training data formats. We found that it is best to input the coordinates of the viewpoints together with the four camera images and to use images from random viewpoints as the training data.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 1","pages":"Pages 1-10"},"PeriodicalIF":0.0,"publicationDate":"2023-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49830406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Metaverse Virtual Social Center for the Elderly Communication During the Social Distancing 虚拟虚拟社交中心在老年人社交距离中的应用
Q1 Computer Science Pub Date : 2023-02-01 DOI: 10.1016/j.vrih.2022.07.007
Hui Liang, Jiupeng Li, Yi Wang, Junjun Pan, Yazhou Zhang, Xiaohang Dong
{"title":"Metaverse Virtual Social Center for the Elderly Communication During the Social Distancing","authors":"Hui Liang, Jiupeng Li, Yi Wang, Junjun Pan, Yazhou Zhang, Xiaohang Dong","doi":"10.1016/j.vrih.2022.07.007","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.07.007","url":null,"abstract":"","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 1","pages":"68 - 80"},"PeriodicalIF":0.0,"publicationDate":"2023-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"55184258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Unrolling Rain-guided Detail Recovery Network for Single Image Deraining 推出用于单图像降阶的雨水引导细节恢复网络
Q1 Computer Science Pub Date : 2023-02-01 DOI: 10.1016/j.vrih.2022.06.002
Kailong Lin, Shaowei Zhang, Yu Luo, Jie Ling

Owing to the rapid development of deep networks, single image deraining tasks have achieved significant progress. Various architectures have been designed to recursively or directly remove rain, and most rain streaks can be removed by existing deraining methods. However, many of them cause a loss of details during deraining, resulting in visual artifacts. To resolve the detail-losing issue, we propose a novel unrolling rain-guided detail recovery network (URDRN) for single image deraining based on the observation that the most degraded areas of the background image tend to be the most rain-corrupted regions. Furthermore, to address the problem that most existing deep-learning-based methods trivialize the observation model and simply learn an end-to-end mapping, the proposed URDRN unrolls the single image deraining task into two subproblems: rain extraction and detail recovery. Specifically, first, a context aggregation attention network is introduced to effectively extract rain streaks, and then, a rain attention map is generated as an indicator to guide the detail-recovery process. For a detail-recovery sub-network, with the guidance of the rain attention map, a simple encoder–decoder model is sufficient to recover the lost details. Experiments on several well-known benchmark datasets show that the proposed approach can achieve a competitive performance in comparison with other state-of-the-art methods.

由于深度网络的快速发展,单图像去噪任务取得了重大进展。已经设计了各种架构来递归地或直接地去除雨水,并且大多数雨条纹可以通过现有的去噪方法来去除。然而,它们中的许多会在去噪过程中导致细节丢失,从而导致视觉伪影。为了解决细节丢失问题,我们提出了一种新的展开雨水引导细节恢复网络(URDRN),用于单图像去噪,该网络基于对背景图像中退化程度最高的区域往往是雨水破坏程度最高的地区的观察。此外,为了解决大多数现有的基于深度学习的方法轻视观测模型并简单地学习端到端映射的问题,所提出的URDRN将单个图像去噪任务分解为两个子问题:雨水提取和细节恢复。具体来说,首先引入上下文聚合注意力网络来有效地提取雨带,然后生成雨带注意力图作为指标来指导细节恢复过程。对于细节恢复子网络,在雨水注意力图的指导下,一个简单的编码器-解码器模型就足以恢复丢失的细节。在几个著名的基准数据集上的实验表明,与其他最先进的方法相比,所提出的方法可以获得有竞争力的性能。
{"title":"Unrolling Rain-guided Detail Recovery Network for Single Image Deraining","authors":"Kailong Lin,&nbsp;Shaowei Zhang,&nbsp;Yu Luo,&nbsp;Jie Ling","doi":"10.1016/j.vrih.2022.06.002","DOIUrl":"https://doi.org/10.1016/j.vrih.2022.06.002","url":null,"abstract":"<div><p>Owing to the rapid development of deep networks, single image deraining tasks have achieved significant progress. Various architectures have been designed to recursively or directly remove rain, and most rain streaks can be removed by existing deraining methods. However, many of them cause a loss of details during deraining, resulting in visual artifacts. To resolve the detail-losing issue, we propose a novel unrolling rain-guided detail recovery network (URDRN) for single image deraining based on the observation that the most degraded areas of the background image tend to be the most rain-corrupted regions. Furthermore, to address the problem that most existing deep-learning-based methods trivialize the observation model and simply learn an end-to-end mapping, the proposed URDRN unrolls the single image deraining task into two subproblems: rain extraction and detail recovery. Specifically, first, a context aggregation attention network is introduced to effectively extract rain streaks, and then, a rain attention map is generated as an indicator to guide the detail-recovery process. For a detail-recovery sub-network, with the guidance of the rain attention map, a simple encoder–decoder model is sufficient to recover the lost details. Experiments on several well-known benchmark datasets show that the proposed approach can achieve a competitive performance in comparison with other state-of-the-art methods.</p></div>","PeriodicalId":33538,"journal":{"name":"Virtual Reality Intelligent Hardware","volume":"5 1","pages":"Pages 11-23"},"PeriodicalIF":0.0,"publicationDate":"2023-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49830408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Virtual Reality Intelligent Hardware
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1