The combination of terahertz and massive multiple-input multiple-output (MIMO) is promising for meeting the increasing data rate demand of future wireless communication systems thanks to the significant bandwidth and spatial degrees of freedom. However, unique channel features, such as the near-field beam split effect, make channel estimation particularly challenging in terahertz massive MIMO systems. On one hand, adopting the conventional angular domain transformation dictionary designed for low-frequency far-field channels will result in degraded channel sparsity and destroyed sparsity structure in the transformed domain. On the other hand, most existing compressive sensing based channel estimation algorithms cannot achieve high performance and low complexity simultaneously. To alleviate these issues, in this study, we first adopt frequency-dependent near-field dictionaries to maintain good channel sparsity and sparsity structure in the transformed domain under the near-field beam split effect. Then, a deep unfolding based wideband terahertz massive MIMO channel estimation algorithm is proposed. In each iteration of the approximate message passing-sparse Bayesian learning algorithm, the optimal update rule is learned by a deep neural network (DNN), whose architecture is customized to effectively exploit the inherent channel patterns. Furthermore, a mixed training method based on novel designs of the DNN architecture and the loss function is developed to effectively train data from different system configurations. Simulation results validate the superiority of the proposed algorithm in terms of performance, complexity, and robustness.
太赫兹和大规模多输入多输出(MIMO)技术的结合具有巨大的带宽和空间自由度,有望满足未来无线通信系统日益增长的数据传输速率需求。然而,近场波束分裂效应等独特的信道特征使得太赫兹大规模多输入多输出系统中的信道估计变得尤为困难。一方面,采用为低频远场信道设计的传统角域变换字典会导致信道稀疏性下降,破坏变换域中的稀疏性结构。另一方面,现有的基于压缩感知的信道估计算法大多无法同时实现高性能和低复杂度。为了解决这些问题,在本研究中,我们首先采用频率相关的近场字典,在近场波束分裂效应下保持变换域中良好的信道稀疏性和稀疏结构。然后,提出了一种基于深度展开的宽带太赫兹大规模 MIMO 信道估计算法。在近似消息传递-稀疏贝叶斯学习算法的每次迭代中,通过深度神经网络(DNN)学习最优更新规则。此外,还开发了一种基于 DNN 架构和损失函数新设计的混合训练方法,以有效训练来自不同系统配置的数据。仿真结果验证了所提算法在性能、复杂性和鲁棒性方面的优越性。
{"title":"Deep unfolding based channel estimation for wideband terahertz near-field massive MIMO systems","authors":"Jiabao Gao, Xiaoming Chen, Geoffrey Ye Li","doi":"10.1631/fitee.2300760","DOIUrl":"https://doi.org/10.1631/fitee.2300760","url":null,"abstract":"<p>The combination of terahertz and massive multiple-input multiple-output (MIMO) is promising for meeting the increasing data rate demand of future wireless communication systems thanks to the significant bandwidth and spatial degrees of freedom. However, unique channel features, such as the near-field beam split effect, make channel estimation particularly challenging in terahertz massive MIMO systems. On one hand, adopting the conventional angular domain transformation dictionary designed for low-frequency far-field channels will result in degraded channel sparsity and destroyed sparsity structure in the transformed domain. On the other hand, most existing compressive sensing based channel estimation algorithms cannot achieve high performance and low complexity simultaneously. To alleviate these issues, in this study, we first adopt frequency-dependent near-field dictionaries to maintain good channel sparsity and sparsity structure in the transformed domain under the near-field beam split effect. Then, a deep unfolding based wideband terahertz massive MIMO channel estimation algorithm is proposed. In each iteration of the approximate message passing-sparse Bayesian learning algorithm, the optimal update rule is learned by a deep neural network (DNN), whose architecture is customized to effectively exploit the inherent channel patterns. Furthermore, a mixed training method based on novel designs of the DNN architecture and the loss function is developed to effectively train data from different system configurations. Simulation results validate the superiority of the proposed algorithm in terms of performance, complexity, and robustness.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"122 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141948095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ziyi Zhou, Chengyue Wang, Kexun Yan, Hui Shi, Xin Pang
Reversible data hiding in encrypted images (RDHEI) is essential for safeguarding sensitive information within the encrypted domain. In this study, we propose an intelligent pixel predictor based on a residual group block and a spatial attention module, showing superior pixel prediction performance compared to existing predictors. Additionally, we introduce an adaptive joint coding method that leverages bit-plane characteristics and intra-block pixel correlations to maximize embedding space, outperforming single coding approaches. The image owner employs the presented intelligent predictor to forecast the original image, followed by encryption through additive secret sharing before conveying the encrypted image to data hiders. Subsequently, data hiders encrypt secret data and embed them within the encrypted image before transmitting the image to the receiver. The receiver can extract secret data and recover the original image losslessly, with the processes of data extraction and image recovery being separable. Our innovative approach combines an intelligent predictor with additive secret sharing, achieving reversible data embedding and extraction while ensuring security and lossless recovery. Experimental results demonstrate that the predictor performs well and has a substantial embedding capacity. For the Lena image, the number of prediction errors within the range of [−5, 5] is as high as 242 500 and our predictor achieves an embedding capacity of 4.39 bpp.
加密图像中的可逆数据隐藏(RDHEI)对于保护加密域内的敏感信息至关重要。在这项研究中,我们提出了一种基于残差组块和空间注意力模块的智能像素预测器,与现有预测器相比,显示出更优越的像素预测性能。此外,我们还引入了一种自适应联合编码方法,该方法利用位平面特征和块内像素相关性最大化嵌入空间,优于单一编码方法。图像所有者利用所介绍的智能预测器预测原始图像,然后通过加法秘密共享进行加密,再将加密图像传送给数据隐藏者。随后,数据隐藏者加密秘密数据并将其嵌入加密图像中,然后再将图像传输给接收者。接收者可以提取秘密数据并无损地恢复原始图像,数据提取和图像恢复过程是可分离的。我们的创新方法将智能预测器与加法秘密共享相结合,实现了可逆数据嵌入和提取,同时确保了安全性和无损恢复。实验结果表明,该预测器性能良好,具有相当大的嵌入能力。对于 Lena 图像,[-5, 5] 范围内的预测错误数高达 242 500,我们的预测器实现了 4.39 bpp 的嵌入容量。
{"title":"Reversible data hiding in encrypted images based on additive secret sharing and additive joint coding using an intelligent predictor","authors":"Ziyi Zhou, Chengyue Wang, Kexun Yan, Hui Shi, Xin Pang","doi":"10.1631/fitee.2300750","DOIUrl":"https://doi.org/10.1631/fitee.2300750","url":null,"abstract":"<p>Reversible data hiding in encrypted images (RDHEI) is essential for safeguarding sensitive information within the encrypted domain. In this study, we propose an intelligent pixel predictor based on a residual group block and a spatial attention module, showing superior pixel prediction performance compared to existing predictors. Additionally, we introduce an adaptive joint coding method that leverages bit-plane characteristics and intra-block pixel correlations to maximize embedding space, outperforming single coding approaches. The image owner employs the presented intelligent predictor to forecast the original image, followed by encryption through additive secret sharing before conveying the encrypted image to data hiders. Subsequently, data hiders encrypt secret data and embed them within the encrypted image before transmitting the image to the receiver. The receiver can extract secret data and recover the original image losslessly, with the processes of data extraction and image recovery being separable. Our innovative approach combines an intelligent predictor with additive secret sharing, achieving reversible data embedding and extraction while ensuring security and lossless recovery. Experimental results demonstrate that the predictor performs well and has a substantial embedding capacity. For the Lena image, the number of prediction errors within the range of [−5, 5] is as high as 242 500 and our predictor achieves an embedding capacity of 4.39 bpp.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"216 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141882833","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
In graphic design, layout is a result of the interaction between the design elements in the foreground and background images. However, prevalent research focuses on enhancing the quality of layout generation algorithms, overlooking the interaction and controllability that are essential for designers when applying these methods in real-world situations. This paper proposes a user-centered layout design system, Iris, which provides designers with an interactive environment to expedite the workflow, and this environment encompasses the features of user-constraint specification, layout generation, custom editing, and final rendering. To satisfy the multiple constraints specified by designers, we introduce a novel generation model, multi-constraint LayoutVQ-VAE, for advancing layout generation under intra- and inter-domain constraints. Qualitative and quantitative experiments on our proposed model indicate that it outperforms or is comparable to prevalent state-of-the-art models in multiple aspects. User studies on Iris further demonstrate that the system significantly enhances design efficiency while achieving human-like layout designs.
{"title":"Iris: a multi-constraint graphic layout generation system","authors":"Liuqing Chen, Qianzhi Jing, Yixin Tsang, Tingting Zhou","doi":"10.1631/fitee.2300312","DOIUrl":"https://doi.org/10.1631/fitee.2300312","url":null,"abstract":"<p>In graphic design, layout is a result of the interaction between the design elements in the foreground and background images. However, prevalent research focuses on enhancing the quality of layout generation algorithms, overlooking the interaction and controllability that are essential for designers when applying these methods in real-world situations. This paper proposes a user-centered layout design system, Iris, which provides designers with an interactive environment to expedite the workflow, and this environment encompasses the features of user-constraint specification, layout generation, custom editing, and final rendering. To satisfy the multiple constraints specified by designers, we introduce a novel generation model, multi-constraint LayoutVQ-VAE, for advancing layout generation under intra- and inter-domain constraints. Qualitative and quantitative experiments on our proposed model indicate that it outperforms or is comparable to prevalent state-of-the-art models in multiple aspects. User studies on Iris further demonstrate that the system significantly enhances design efficiency while achieving human-like layout designs.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"50 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141775015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
As one of the essential tools for spatio–temporal traffic data mining, vehicle trajectory clustering is widely used to mine the behavior patterns of vehicles. However, uploading original vehicle trajectory data to the server and clustering carry the risk of privacy leakage. Therefore, one of the current challenges is determining how to perform vehicle trajectory clustering while protecting user privacy. We propose a privacy-preserving vehicle trajectory clustering framework and construct a vehicle trajectory clustering model (IKV) based on the variational autoencoder (VAE) and an improved K-means algorithm. In the framework, the client calculates the hidden variables of the vehicle trajectory and uploads the variables to the server; the server uses the hidden variables for clustering analysis and delivers the analysis results to the client. The IKV’ workflow is as follows: first, we train the VAE with historical vehicle trajectory data (when VAE’s decoder can approximate the original data, the encoder is deployed to the edge computing device); second, the edge device transmits the hidden variables to the server; finally, clustering is performed using improved K-means, which prevents the leakage of the vehicle trajectory. IKV is compared to numerous clustering methods on three datasets. In the nine performance comparison experiments, IKV achieves optimal or sub-optimal performance in six of the experiments. Furthermore, in the nine sensitivity analysis experiments, IKV not only demonstrates significant stability in seven experiments but also shows good robustness to hyperparameter variations. These results validate that the framework proposed in this paper is not only suitable for privacy-conscious production environments, such as carpooling tasks, but also adapts to clustering tasks of different magnitudes due to the low sensitivity to the number of cluster centers.
{"title":"A privacy-preserving vehicle trajectory clustering framework","authors":"Ran Tian, Pulun Gao, Yanxing Liu","doi":"10.1631/fitee.2300369","DOIUrl":"https://doi.org/10.1631/fitee.2300369","url":null,"abstract":"<p>As one of the essential tools for spatio–temporal traffic data mining, vehicle trajectory clustering is widely used to mine the behavior patterns of vehicles. However, uploading original vehicle trajectory data to the server and clustering carry the risk of privacy leakage. Therefore, one of the current challenges is determining how to perform vehicle trajectory clustering while protecting user privacy. We propose a privacy-preserving vehicle trajectory clustering framework and construct a vehicle trajectory clustering model (IKV) based on the variational autoencoder (VAE) and an improved <i>K</i>-means algorithm. In the framework, the client calculates the hidden variables of the vehicle trajectory and uploads the variables to the server; the server uses the hidden variables for clustering analysis and delivers the analysis results to the client. The IKV’ workflow is as follows: first, we train the VAE with historical vehicle trajectory data (when VAE’s decoder can approximate the original data, the encoder is deployed to the edge computing device); second, the edge device transmits the hidden variables to the server; finally, clustering is performed using improved <i>K</i>-means, which prevents the leakage of the vehicle trajectory. IKV is compared to numerous clustering methods on three datasets. In the nine performance comparison experiments, IKV achieves optimal or sub-optimal performance in six of the experiments. Furthermore, in the nine sensitivity analysis experiments, IKV not only demonstrates significant stability in seven experiments but also shows good robustness to hyperparameter variations. These results validate that the framework proposed in this paper is not only suitable for privacy-conscious production environments, such as carpooling tasks, but also adapts to clustering tasks of different magnitudes due to the low sensitivity to the number of cluster centers.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"81 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141775016","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The integration of industrial Internet, cloud computing, and big data technology is changing the business and management mode of the industry chain. However, the industry chain is characterized by a wide range of fields, complex environment, and many factors, which creates a challenge for efficient integration and leveraging of industrial big data. Aiming at the integration of physical space and virtual space of the current industry chain, we propose an industry chain digital twin (DT) system framework for the industrial Internet. In addition, an industry chain information model based on a knowledge graph (KG) is proposed to integrate complex and heterogeneous industry chain data and extract industrial knowledge. First, the ontology of the industry chain is established, and an entity alignment method based on scientific and technological achievements is proposed. Second, the bidirectional encoder representations from Transformers (BERT) based multi-head selection model is proposed for joint entity–relation extraction of industry chain information. Third, a relation completion model based on a relational graph convolutional network (R-GCN) and a graph sample and aggregate network (GraphSAGE) is proposed which considers both semantic information and graph structure information of KG. Experimental results show that the performances of the proposed joint entity–relation extraction model and relation completion model are significantly better than those of the baselines. Finally, an industry chain information model is established based on the data of 18 industry chains in the field of basic machinery, which proves the feasibility of the proposed method.
工业互联网、云计算、大数据技术的融合正在改变产业链的业务和管理模式。然而,产业链领域广泛、环境复杂、因素众多,给工业大数据的高效整合和利用带来了挑战。针对当前产业链物理空间与虚拟空间的融合,我们提出了面向工业互联网的产业链数字孪生(DT)系统框架。此外,还提出了基于知识图谱(KG)的产业链信息模型,以整合复杂的异构产业链数据并提取产业知识。首先,建立了产业链本体,并提出了基于科技成果的实体对齐方法。其次,提出了基于变压器双向编码器表示(BERT)的多头选择模型,用于产业链信息的联合实体-关系提取。第三,提出了基于关系图卷积网络(R-GCN)和图样本与聚合网络(GraphSAGE)的关系完成模型,该模型同时考虑了 KG 的语义信息和图结构信息。实验结果表明,所提出的联合实体-关系提取模型和关系完成模型的性能明显优于基线模型。最后,基于基础机械领域 18 条产业链的数据建立了产业链信息模型,证明了所提方法的可行性。
{"title":"Digital twin system framework and information model for industry chain based on industrial Internet","authors":"Wenxuan Wang, Yongqin Liu, Xudong Chai, Lin Zhang","doi":"10.1631/fitee.2300123","DOIUrl":"https://doi.org/10.1631/fitee.2300123","url":null,"abstract":"<p>The integration of industrial Internet, cloud computing, and big data technology is changing the business and management mode of the industry chain. However, the industry chain is characterized by a wide range of fields, complex environment, and many factors, which creates a challenge for efficient integration and leveraging of industrial big data. Aiming at the integration of physical space and virtual space of the current industry chain, we propose an industry chain digital twin (DT) system framework for the industrial Internet. In addition, an industry chain information model based on a knowledge graph (KG) is proposed to integrate complex and heterogeneous industry chain data and extract industrial knowledge. First, the ontology of the industry chain is established, and an entity alignment method based on scientific and technological achievements is proposed. Second, the bidirectional encoder representations from Transformers (BERT) based multi-head selection model is proposed for joint entity–relation extraction of industry chain information. Third, a relation completion model based on a relational graph convolutional network (R-GCN) and a graph sample and aggregate network (GraphSAGE) is proposed which considers both semantic information and graph structure information of KG. Experimental results show that the performances of the proposed joint entity–relation extraction model and relation completion model are significantly better than those of the baselines. Finally, an industry chain information model is established based on the data of 18 industry chains in the field of basic machinery, which proves the feasibility of the proposed method.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"40 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141775014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The game of Tibetan Go faces the scarcity of expert knowledge and research literature. Therefore, we study the zero learning model of Tibetan Go under limited computing power resources and propose a novel scale-invariant U-Net style two-headed output lightweight network TibetanGoTinyNet. The lightweight convolutional neural networks and capsule structure are applied to the encoder and decoder of TibetanGoTinyNet to reduce computational burden and achieve better feature extraction results. Several autonomous self-attention mechanisms are integrated into TibetanGoTinyNet to capture the Tibetan Go board’s spatial and global information and select important channels. The training data are generated entirely from self-play games. TibetanGoTinyNet achieves 62%–78% winning rate against other four U-Net style models including Res-UNet, Res-UNet Attention, Ghost-UNet, and Ghost Capsule-UNet. It also achieves 75% winning rate in the ablation experiments on the attention mechanism with embedded positional information. The model saves about 33% of the training time with 45%–50% winning rate for different Monte-Carlo tree search (MCTS) simulation counts when migrated from 9 × 9 to 11 × 11 boards. Code for our model is available at https://github.com/paulzyy/TibetanGoTinyNet.
{"title":"TibetanGoTinyNet: a lightweight U-Net style network for zero learning of Tibetan Go","authors":"Xiali Li, Yanyin Zhang, Licheng Wu, Yandong Chen, Junzhi Yu","doi":"10.1631/fitee.2300493","DOIUrl":"https://doi.org/10.1631/fitee.2300493","url":null,"abstract":"<p>The game of Tibetan Go faces the scarcity of expert knowledge and research literature. Therefore, we study the zero learning model of Tibetan Go under limited computing power resources and propose a novel scale-invariant U-Net style two-headed output lightweight network TibetanGoTinyNet. The lightweight convolutional neural networks and capsule structure are applied to the encoder and decoder of TibetanGoTinyNet to reduce computational burden and achieve better feature extraction results. Several autonomous self-attention mechanisms are integrated into TibetanGoTinyNet to capture the Tibetan Go board’s spatial and global information and select important channels. The training data are generated entirely from self-play games. TibetanGoTinyNet achieves 62%–78% winning rate against other four U-Net style models including Res-UNet, Res-UNet Attention, Ghost-UNet, and Ghost Capsule-UNet. It also achieves 75% winning rate in the ablation experiments on the attention mechanism with embedded positional information. The model saves about 33% of the training time with 45%–50% winning rate for different Monte-Carlo tree search (MCTS) simulation counts when migrated from 9 × 9 to 11 × 11 boards. Code for our model is available at https://github.com/paulzyy/TibetanGoTinyNet.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"29 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141785342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Ground elevation estimation is vital for numerous applications in autonomous vehicles and intelligent robotics including three-dimensional object detection, navigable space detection, point cloud matching for localization, and registration for mapping. However, most works regard the ground as a plane without height information, which causes inaccurate manipulation in these applications. In this work, we propose GeeNet, a novel end-to-end, lightweight method that completes the ground in nearly real time and simultaneously estimates the ground elevation in a grid-based representation. GeeNet leverages the mixing of two- and three-dimensional convolutions to preserve a lightweight architecture to regress ground elevation information for each cell of the grid. For the first time, GeeNet has fulfilled ground elevation estimation from semantic scene completion. We use the SemanticKITTI and SemanticPOSS datasets to validate the proposed GeeNet, demonstrating the qualitative and quantitative performances of GeeNet on ground elevation estimation and semantic scene completion of the point cloud. Moreover, the cross-dataset generalization capability of GeeNet is experimentally proven. GeeNet achieves state-of-the-art performance in terms of point cloud completion and ground elevation estimation, with a runtime of 0.88 ms.
{"title":"GeeNet: robust and fast point cloud completion for ground elevation estimation towards autonomous vehicles","authors":"Liwen Liu, Weidong Yang, Ben Fei","doi":"10.1631/fitee.2300388","DOIUrl":"https://doi.org/10.1631/fitee.2300388","url":null,"abstract":"<p>Ground elevation estimation is vital for numerous applications in autonomous vehicles and intelligent robotics including three-dimensional object detection, navigable space detection, point cloud matching for localization, and registration for mapping. However, most works regard the ground as a plane without height information, which causes inaccurate manipulation in these applications. In this work, we propose GeeNet, a novel end-to-end, lightweight method that completes the ground in nearly real time and simultaneously estimates the ground elevation in a grid-based representation. GeeNet leverages the mixing of two- and three-dimensional convolutions to preserve a lightweight architecture to regress ground elevation information for each cell of the grid. For the first time, GeeNet has fulfilled ground elevation estimation from semantic scene completion. We use the SemanticKITTI and SemanticPOSS datasets to validate the proposed GeeNet, demonstrating the qualitative and quantitative performances of GeeNet on ground elevation estimation and semantic scene completion of the point cloud. Moreover, the cross-dataset generalization capability of GeeNet is experimentally proven. GeeNet achieves state-of-the-art performance in terms of point cloud completion and ground elevation estimation, with a runtime of 0.88 ms.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"67 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141775129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Lingjing Li, Chunyang Ma, Nian Zhao, Jie Peng, Bin Liu, Haining Ji, Yuchen Wang, Pinghua Tang
Dy3+-doped fluoride fiber lasers have important applications in environment monitoring, real-time sensing, and polymer processing. At present, achieving a high-efficiency and high-power Dy3+-doped fluoride fiber laser in the mid-infrared (mid-IR) region over 3 µm is a scientific and technological frontier. Typically, Dy3+-doped fluoride fiber lasers use a unidirectional pumping method, which suffers from the drawback of high thermal loading density on the fiber tips, thus limiting power scalability. In this study, a bi-directional in-band pumping scheme, to address the limitations of output power scaling and to enhance the efficiency of the Dy3+-doped fluoride fiber laser at 3.2 µm, is investigated numerically based on rate equations and propagation equations. Detailed simulation results reveal that the optical–optical efficiency of the bi-directional in-band pumped Dy3+-doped fluoride fiber laser can reach 75.1%, approaching the Stokes limit of 87.3%. The potential for further improvement of the efficiency of the Dy3+-doped fluoride fiber laser is also discussed. The bi-directional pumping scheme offers the intrinsic advantage of mitigating the thermal load on the fiber tips, unlike unidirectional pumping, in addition to its high efficiency. As a result, it is expected to significantly scale the power output of Dy3+-doped fluoride fiber lasers in the mid-IR regime.
{"title":"Numerical study of a bi-directional in-band pumped dysprosium-doped fluoride fiber laser at 3.2 µm","authors":"Lingjing Li, Chunyang Ma, Nian Zhao, Jie Peng, Bin Liu, Haining Ji, Yuchen Wang, Pinghua Tang","doi":"10.1631/fitee.2300701","DOIUrl":"https://doi.org/10.1631/fitee.2300701","url":null,"abstract":"<p>Dy<sup>3+</sup>-doped fluoride fiber lasers have important applications in environment monitoring, real-time sensing, and polymer processing. At present, achieving a high-efficiency and high-power Dy<sup>3+</sup>-doped fluoride fiber laser in the mid-infrared (mid-IR) region over 3 µm is a scientific and technological frontier. Typically, Dy<sup>3+</sup>-doped fluoride fiber lasers use a unidirectional pumping method, which suffers from the drawback of high thermal loading density on the fiber tips, thus limiting power scalability. In this study, a bi-directional in-band pumping scheme, to address the limitations of output power scaling and to enhance the efficiency of the Dy<sup>3+</sup>-doped fluoride fiber laser at 3.2 µm, is investigated numerically based on rate equations and propagation equations. Detailed simulation results reveal that the optical–optical efficiency of the bi-directional in-band pumped Dy<sup>3+</sup>-doped fluoride fiber laser can reach 75.1%, approaching the Stokes limit of 87.3%. The potential for further improvement of the efficiency of the Dy<sup>3+</sup>-doped fluoride fiber laser is also discussed. The bi-directional pumping scheme offers the intrinsic advantage of mitigating the thermal load on the fiber tips, unlike unidirectional pumping, in addition to its high efficiency. As a result, it is expected to significantly scale the power output of Dy<sup>3+</sup>-doped fluoride fiber lasers in the mid-IR regime.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"61 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141785344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Currently, decarbonization has become an emerging trend in the power system arena. However, the increasing number of photovoltaic units distributed into a distribution network may result in voltage issues, providing challenges for voltage regulation across a large-scale power grid network. Reinforcement learning based intelligent control of smart inverters and other smart building energy management (EM) systems can be leveraged to alleviate these issues. To achieve the best EM strategy for building microgrids in a power system, this paper presents two large-scale multi-agent strategy evaluation methods to preserve building occupants’ comfort while pursuing system-level objectives. The EM problem is formulated as a general-sum game to optimize the benefits at both the system and building levels. The α-rank algorithm can solve the general-sum game and guarantee the ranking theoretically, but it is limited by the interaction complexity and hardly applies to the practical power system. A new evaluation algorithm (TcEval) is proposed by practically scaling the α-rank algorithm through a tensor complement to reduce the interaction complexity. Then, considering the noise prevalent in practice, a noise processing model with domain knowledge is built to calculate the strategy payoffs, and thus the TcEval-AS algorithm is proposed when noise exists. Both evaluation algorithms developed in this paper greatly reduce the interaction complexity compared with existing approaches, including ResponseGraphUCB (RG-UCB) and αInformationGain (α-IG). Finally, the effectiveness of the proposed algorithms is verified in the EM case with realistic data.
目前,去碳化已成为电力系统领域的新兴趋势。然而,分布在配电网络中的光伏装置数量不断增加,可能会导致电压问题,给大规模电网网络的电压调节带来挑战。基于强化学习的智能逆变器和其他智能建筑能源管理(EM)系统的智能控制可以有效缓解这些问题。为了实现电力系统中楼宇微电网的最佳 EM 策略,本文提出了两种大规模多代理策略评估方法,以在追求系统级目标的同时保持楼宇居住者的舒适度。电磁问题被表述为一般和博弈,以优化系统和楼宇两个层面的效益。α-rank算法可以求解泛和博弈并保证理论上的排序,但受限于交互复杂性,很难应用于实际的电力系统。本文提出了一种新的评估算法(TcEval),通过张量补法对α-rank 算法进行实际扩展,以降低交互复杂度。然后,考虑到实践中普遍存在的噪声,建立了一个具有领域知识的噪声处理模型来计算策略回报,从而提出了存在噪声时的 TcEval-AS 算法。与现有方法(包括 ResponseGraphUCB (RG-UCB) 和 αInformationGain (α-IG))相比,本文开发的两种评估算法都大大降低了交互复杂度。最后,在 EM 案例中用现实数据验证了所提算法的有效性。
{"title":"Multi-agent evaluation for energy management by practically scaling α-rank","authors":"Yiyun Sun, Senlin Zhang, Meiqin Liu, Ronghao Zheng, Shanling Dong, Xuguang Lan","doi":"10.1631/fitee.2300438","DOIUrl":"https://doi.org/10.1631/fitee.2300438","url":null,"abstract":"<p>Currently, decarbonization has become an emerging trend in the power system arena. However, the increasing number of photovoltaic units distributed into a distribution network may result in voltage issues, providing challenges for voltage regulation across a large-scale power grid network. Reinforcement learning based intelligent control of smart inverters and other smart building energy management (EM) systems can be leveraged to alleviate these issues. To achieve the best EM strategy for building microgrids in a power system, this paper presents two large-scale multi-agent strategy evaluation methods to preserve building occupants’ comfort while pursuing system-level objectives. The EM problem is formulated as a general-sum game to optimize the benefits at both the system and building levels. The <i>α</i>-rank algorithm can solve the general-sum game and guarantee the ranking theoretically, but it is limited by the interaction complexity and hardly applies to the practical power system. A new evaluation algorithm (TcEval) is proposed by practically scaling the <i>α</i>-rank algorithm through a tensor complement to reduce the interaction complexity. Then, considering the noise prevalent in practice, a noise processing model with domain knowledge is built to calculate the strategy payoffs, and thus the TcEval-AS algorithm is proposed when noise exists. Both evaluation algorithms developed in this paper greatly reduce the interaction complexity compared with existing approaches, including ResponseGraphUCB (RG-UCB) and <i>α</i>InformationGain (<i>α</i>-IG). Finally, the effectiveness of the proposed algorithms is verified in the EM case with realistic data.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"12 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141785343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper investigates the issue of adaptive optimal tracking control for nonlinear systems with dynamic state constraints. An asymmetric time-varying integral barrier Lyapunov function (ATIBLF) based integral reinforcement learning (IRL) control algorithm with an actor–critic structure is first proposed. The ATIBLF items are appropriately arranged in every step of the optimized backstepping control design to ensure that the dynamic full-state constraints are never violated. Thus, optimal virtual/actual control in every backstepping subsystem is decomposed with ATIBLF items and also with an adaptive optimized item. Meanwhile, neural networks are used to approximate the gradient value functions. According to the Lyapunov stability theorem, the boundedness of all signals of the closed-loop system is proved, and the proposed control scheme ensures that the system states are within predefined compact sets. Finally, the effectiveness of the proposed control approach is validated by simulations.
{"title":"Asymmetric time-varying integral barrier Lyapunov function based adaptive optimal control for nonlinear systems with dynamic state constraints","authors":"Yan Wei, Mingshuang Hao, Xinyi Yu, Linlin Ou","doi":"10.1631/fitee.2300675","DOIUrl":"https://doi.org/10.1631/fitee.2300675","url":null,"abstract":"<p>This paper investigates the issue of adaptive optimal tracking control for nonlinear systems with dynamic state constraints. An asymmetric time-varying integral barrier Lyapunov function (ATIBLF) based integral reinforcement learning (IRL) control algorithm with an actor–critic structure is first proposed. The ATIBLF items are appropriately arranged in every step of the optimized backstepping control design to ensure that the dynamic full-state constraints are never violated. Thus, optimal virtual/actual control in every backstepping subsystem is decomposed with ATIBLF items and also with an adaptive optimized item. Meanwhile, neural networks are used to approximate the gradient value functions. According to the Lyapunov stability theorem, the boundedness of all signals of the closed-loop system is proved, and the proposed control scheme ensures that the system states are within predefined compact sets. Finally, the effectiveness of the proposed control approach is validated by simulations.</p>","PeriodicalId":12608,"journal":{"name":"Frontiers of Information Technology & Electronic Engineering","volume":"60 1","pages":""},"PeriodicalIF":3.0,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141548128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}