首页 > 最新文献

International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)最新文献

英文 中文
Exploiting the power of StarGANv2 in the wild 在野外利用StarGANv2的力量
Gengjun Huang, Xiaosheng Long, Yiming Mao
With wide-spread usage of style transfer, numerous methods for style transfer draw an increasing attention. Several methods to enhance the efficiency of style transformers have been made, one of them is StarGANv2, a method for multiple-style transfer, which can transform a batch of source pictures into other pictures with different styles. The main difference of StarGANv2 with other style transformers is that it uses style code to represent the styles to enable StarGANv2 to complete multiple-style transformation. The authors of StarGANv2 use CelebA-HQ and AFHQ dataset to train the model and test the model, and the results are pretty better than other style transformers. The goal of this paper is to exploit the effectiveness of StarGANv2 in the real-world scenes, such as over exposure or the angle facing the camera. The results validate the power of StarGANv2 where the model is robust enough to transfer the pictures into other styles. To achieve this, the authors of StarGANv2 use the photo clipped in videos which record real-world animals and form a new dataset. Then, the authors of StarGANv2 use the dataset to test the pre-trained model which is trained by AFHQ dataset and evaluate it according to FID metric. The authors of StarGANv2 draw a conclusion that StarGANv2 is robust in real world scenes. The meaning of this paper is that the authors get the real-world usage of StarGANv2 and have a test of StarGANv2’s robustness in real world photos and validate the potential of StarGANv2 in real-world applications.
随着风格迁移的广泛使用,许多风格迁移的方法越来越受到人们的关注。提出了几种提高样式转换效率的方法,其中一种是StarGANv2,它是一种多样式转换方法,可以将一批源图片转换为具有不同样式的其他图片。StarGANv2与其他样式转换器的主要区别在于,它使用样式代码来表示样式,使StarGANv2能够完成多样式转换。StarGANv2的作者使用CelebA-HQ和AFHQ数据集对模型进行训练和测试,结果比其他风格变形器要好得多。本文的目标是利用StarGANv2在现实世界场景中的有效性,例如过度曝光或面向相机的角度。结果验证了StarGANv2的强大功能,该模型具有足够的鲁棒性,可以将图像转换为其他样式。为了实现这一目标,StarGANv2的作者使用了记录现实世界动物的视频剪辑的照片,并形成了一个新的数据集。然后,StarGANv2使用该数据集对AFHQ数据集训练的预训练模型进行测试,并根据FID度量对其进行评估。StarGANv2的作者得出结论,StarGANv2在现实世界场景中是健壮的。本文的意义在于,作者获得了StarGANv2在现实世界中的使用情况,并在真实世界的照片中测试了StarGANv2的鲁棒性,验证了StarGANv2在现实应用中的潜力。
{"title":"Exploiting the power of StarGANv2 in the wild","authors":"Gengjun Huang, Xiaosheng Long, Yiming Mao","doi":"10.1117/12.2671374","DOIUrl":"https://doi.org/10.1117/12.2671374","url":null,"abstract":"With wide-spread usage of style transfer, numerous methods for style transfer draw an increasing attention. Several methods to enhance the efficiency of style transformers have been made, one of them is StarGANv2, a method for multiple-style transfer, which can transform a batch of source pictures into other pictures with different styles. The main difference of StarGANv2 with other style transformers is that it uses style code to represent the styles to enable StarGANv2 to complete multiple-style transformation. The authors of StarGANv2 use CelebA-HQ and AFHQ dataset to train the model and test the model, and the results are pretty better than other style transformers. The goal of this paper is to exploit the effectiveness of StarGANv2 in the real-world scenes, such as over exposure or the angle facing the camera. The results validate the power of StarGANv2 where the model is robust enough to transfer the pictures into other styles. To achieve this, the authors of StarGANv2 use the photo clipped in videos which record real-world animals and form a new dataset. Then, the authors of StarGANv2 use the dataset to test the pre-trained model which is trained by AFHQ dataset and evaluate it according to FID metric. The authors of StarGANv2 draw a conclusion that StarGANv2 is robust in real world scenes. The meaning of this paper is that the authors get the real-world usage of StarGANv2 and have a test of StarGANv2’s robustness in real world photos and validate the potential of StarGANv2 in real-world applications.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127453153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Human-object interaction detection based on graph model 基于图模型的人-物交互检测
Qing Ye, Xiujuan Xu
Human-Object Interaction (HOI) detection is a fundamental task for understanding real-world scenes. In this paper, a graph model-based human-object interaction detection algorithm is proposed, which aims to make full use of the visual-spatial features and semantic information of human-object instances in the image, thereby improving the accuracy of interaction detection. Aiming at the characteristics of visual-spatial features and semantic information, we take the visual features of human and object instance boxes as nodes, and the corresponding spatial features of interaction relations as edges to construct an initial dense graph, and adaptively update the graph through the spatial and semantic information of instances. The V-COCO dataset is used to evaluate the algorithm, and the final accuracy is significantly improved, which proves the effectiveness of the algorithm.
人-物交互(HOI)检测是理解现实世界场景的一项基本任务。本文提出了一种基于图模型的人-物交互检测算法,该算法旨在充分利用图像中人-物实例的视觉空间特征和语义信息,从而提高交互检测的准确性。针对视觉空间特征和语义信息的特点,以人与物实例盒的视觉特征为节点,以交互关系对应的空间特征为边,构造初始密集图,并通过实例的空间和语义信息自适应更新图。利用V-COCO数据集对算法进行评估,最终的准确率有了明显提高,证明了算法的有效性。
{"title":"Human-object interaction detection based on graph model","authors":"Qing Ye, Xiujuan Xu","doi":"10.1117/12.2671248","DOIUrl":"https://doi.org/10.1117/12.2671248","url":null,"abstract":"Human-Object Interaction (HOI) detection is a fundamental task for understanding real-world scenes. In this paper, a graph model-based human-object interaction detection algorithm is proposed, which aims to make full use of the visual-spatial features and semantic information of human-object instances in the image, thereby improving the accuracy of interaction detection. Aiming at the characteristics of visual-spatial features and semantic information, we take the visual features of human and object instance boxes as nodes, and the corresponding spatial features of interaction relations as edges to construct an initial dense graph, and adaptively update the graph through the spatial and semantic information of instances. The V-COCO dataset is used to evaluate the algorithm, and the final accuracy is significantly improved, which proves the effectiveness of the algorithm.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126913547","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Mural image shedding diseases inpainting algorithm based on structure priority 基于结构优先级的壁画图像脱病算法
Haibo Pen, Shuangshuang Wang, Zhuofan Zhang
The painted murals in Mogao Grottoes and Longmen Grottoes are symbols of China history and culture. However, most of the murals with complex texture and structure have suffered from different degrees of disease erosion after thousands of years. It is necessary to restore the damaged parts of the murals and to accurately restore their contents. In recent years, the use of new virtual technologies such as digital images to repair the damage can largely avoid secondary damage to the murals caused by manual restoration methods. Therefore, this paper takes the restoration of the most typical shedding diseases to the Mogao Caves murals in Dunhuang as an example. Furthermore, the research object of this paper is the shedding diseases including contour lines. For the traditional virtual methods of repairing shedding diseases, the structure and texture are usually restored at the same time, and these methods have little effect on the accurate removal of shedding disease through the contour line. It can be seen that shedding disease through the contour line is more difficult to repair, and more appropriate inpainting methods need to be explored. Considering the particularity of the shedding disease that passes through the contour line, this paper proposes a mural image inpainting algorithm based on structure priority to repair the shedding diseases. First, the structure repair problem is further converted into a optimization problem, and then the global optimization capability of the genetic algorithm is used to realize the connection of the structure information of the damaged area. Then, the texture is filled by subarea optimization to obtain an ideal repair effect, which can reasonably and effectively solve the problem of shedding disease repair through the contour line. Subjective and objective evaluation of experimental results is also better than other comparative methods.
莫高窟和龙门石窟的壁画是中国历史文化的象征。然而,大多数纹理结构复杂的壁画,经过数千年的发展,都遭受了不同程度的疾病侵蚀。有必要修复壁画受损的部分,并准确地恢复壁画的内容。近年来,利用数字图像等新型虚拟技术对壁画进行修复,可以在很大程度上避免手工修复方法对壁画造成的二次损坏。因此,本文以敦煌莫高窟壁画最典型的脱落病修复为例。此外,本文的研究对象是包括等高线在内的脱毛病。对于传统的脱毛病修复虚拟方法,通常是同时修复结构和纹理,这些方法对于通过轮廓线精确去除脱毛病的效果不大。由此可见,通过等高线脱落的疾病更难修复,需要探索更合适的补漆方法。考虑到脱落病通过等高线的特殊性,提出了一种基于结构优先级的壁画图像修复算法。首先将结构修复问题进一步转化为优化问题,然后利用遗传算法的全局优化能力实现受损区域结构信息的连接。然后对纹理进行分区优化填充,获得理想的修复效果,可以合理有效地解决通过轮廓线进行脱落病修复的问题。实验结果的主客观评价也优于其他比较方法。
{"title":"Mural image shedding diseases inpainting algorithm based on structure priority","authors":"Haibo Pen, Shuangshuang Wang, Zhuofan Zhang","doi":"10.1117/12.2671230","DOIUrl":"https://doi.org/10.1117/12.2671230","url":null,"abstract":"The painted murals in Mogao Grottoes and Longmen Grottoes are symbols of China history and culture. However, most of the murals with complex texture and structure have suffered from different degrees of disease erosion after thousands of years. It is necessary to restore the damaged parts of the murals and to accurately restore their contents. In recent years, the use of new virtual technologies such as digital images to repair the damage can largely avoid secondary damage to the murals caused by manual restoration methods. Therefore, this paper takes the restoration of the most typical shedding diseases to the Mogao Caves murals in Dunhuang as an example. Furthermore, the research object of this paper is the shedding diseases including contour lines. For the traditional virtual methods of repairing shedding diseases, the structure and texture are usually restored at the same time, and these methods have little effect on the accurate removal of shedding disease through the contour line. It can be seen that shedding disease through the contour line is more difficult to repair, and more appropriate inpainting methods need to be explored. Considering the particularity of the shedding disease that passes through the contour line, this paper proposes a mural image inpainting algorithm based on structure priority to repair the shedding diseases. First, the structure repair problem is further converted into a optimization problem, and then the global optimization capability of the genetic algorithm is used to realize the connection of the structure information of the damaged area. Then, the texture is filled by subarea optimization to obtain an ideal repair effect, which can reasonably and effectively solve the problem of shedding disease repair through the contour line. Subjective and objective evaluation of experimental results is also better than other comparative methods.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126191158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A multi-scale sarcasm sentiment recognition algorithm incorporating sentence hierarchical representation 基于句子层次表示的多尺度讽刺情感识别算法
Yurong Hao, Long Zhang, Qiusheng Zheng, Liyue Niu
Sarcasm is a special kind of linguistic sentiment that is widely used in a wide range of social media to express strong emotions in users. Therefore, the task of sarcasm recognition is particularly important for social media analysis. There are few studies on sarcasm sentiment recognition in Chinese, and they often ignore the complex interactions between different syntactic components of a sentence, such as sentiment words, entities, dummy words, and special punctuation that occur in the text. In order to improve the accuracy of Chinese sarcasm recognition, this paper proposes a multi-scale neural network sarcasm recognition algorithm incorporating a hierarchical representation of sentences, taking into account the semantic information of sentences and the relationship features between different syntactic components. The hierarchical syntactic tree is reconstructed to distinguish the key components of the sentence, and the multi-channel convolutional network is used to mine the relational features between syntactic levels and deeply fuse them with semantic information to perform the Chinese sarcastic sentiment recognition task. We have tested the method on a publicly available Chinese sarcastic comment dataset, and the results show that the method can effectively improve the accuracy rate of Chinese sarcastic sentiment recognition.
讽刺是一种特殊的语言情感,广泛应用于各种社交媒体中,用来表达用户强烈的情感。因此,讽刺识别的任务对于社交媒体分析尤为重要。汉语讽刺情感识别的研究很少,往往忽略了句子中不同句法成分之间复杂的相互作用,如情感词、实体、假词和文本中出现的特殊标点符号。为了提高汉语讽刺语识别的准确率,本文提出了一种考虑句子语义信息和不同句法成分之间关系特征的句子分层表示的多尺度神经网络讽刺语识别算法。通过重构分层句法树来区分句子的关键成分,利用多通道卷积网络挖掘句法层之间的关系特征,并将其与语义信息深度融合,完成汉语讽刺情感识别任务。我们在一个公开的中文讽刺评论数据集上对该方法进行了测试,结果表明该方法可以有效地提高中文讽刺情感识别的准确率。
{"title":"A multi-scale sarcasm sentiment recognition algorithm incorporating sentence hierarchical representation","authors":"Yurong Hao, Long Zhang, Qiusheng Zheng, Liyue Niu","doi":"10.1117/12.2671064","DOIUrl":"https://doi.org/10.1117/12.2671064","url":null,"abstract":"Sarcasm is a special kind of linguistic sentiment that is widely used in a wide range of social media to express strong emotions in users. Therefore, the task of sarcasm recognition is particularly important for social media analysis. There are few studies on sarcasm sentiment recognition in Chinese, and they often ignore the complex interactions between different syntactic components of a sentence, such as sentiment words, entities, dummy words, and special punctuation that occur in the text. In order to improve the accuracy of Chinese sarcasm recognition, this paper proposes a multi-scale neural network sarcasm recognition algorithm incorporating a hierarchical representation of sentences, taking into account the semantic information of sentences and the relationship features between different syntactic components. The hierarchical syntactic tree is reconstructed to distinguish the key components of the sentence, and the multi-channel convolutional network is used to mine the relational features between syntactic levels and deeply fuse them with semantic information to perform the Chinese sarcastic sentiment recognition task. We have tested the method on a publicly available Chinese sarcastic comment dataset, and the results show that the method can effectively improve the accuracy rate of Chinese sarcastic sentiment recognition.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126069686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparison and analysis of computer vision models based on images of catamount and canid 猫类与犬科动物图像计算机视觉模型的比较与分析
Feng Jiang, Yueyufei Ma, Langyue Wang
Nowadays, target recognition, driverless, medical impact diagnosis, and other applications based on image recognition in life, scientific research, and work, rely mainly on a variety of large models with excellent performance, from the Convolutional Neural Network (CNN) at the beginning to the various variants of the classical model proposed now. In this paper, we will take the example of identifying catamount and canid datasets, comparing the efficiency and accuracy of CNN, Vision Transformer (ViT), and Swin Transformer laterally. We plan to run 25 epochs for each model and record the accuracy and time consumption separately. After the experiments we find that from the comparison of the epoch numbers and the real-time consumption, the CNN takes the least total time, followed by Swin Transformer. Also, ViT takes the least time to reach convergence, while Swin Transformer takes the most time. In terms of training accuracy, ViT has the highest training accuracy, followed by Swin Transformer, and CNN has the lowest training accuracy; the validation accuracy is similar to the training accuracy. ViT has the highest accuracy, but takes the longest time; conversely, CNN takes the shortest time and has the lowest accuracy. Swin Transformer, which seems a combination of CNN and ViT, is most complex but with ideal performance. In the future, ViT is indeed a promising model that deserves further research and exploration to contribute to the computer vision field.
如今,基于图像识别的目标识别、无人驾驶、医疗影响诊断等在生活、科研和工作中的应用,主要依赖于各种性能优异的大型模型,从最开始的卷积神经网络(CNN)到现在提出的经典模型的各种变体。本文将以识别猫和狗的数据集为例,横向比较CNN、Vision Transformer (ViT)和Swin Transformer的效率和准确性。我们计划为每个模型运行25个epoch,并分别记录准确率和耗时。实验发现,从历元数和实时消耗的比较来看,CNN的总时间最少,其次是Swin Transformer。此外,ViT需要最少的时间来达到收敛,而Swin Transformer需要最多的时间。在训练精度方面,ViT的训练精度最高,Swin Transformer次之,CNN的训练精度最低;验证精度与训练精度相近。ViT精度最高,但耗时最长;相反,CNN耗时最短,准确率最低。Swin Transformer看起来是CNN和ViT的结合,它是最复杂的,但性能却很理想。在未来,ViT确实是一个很有前途的模型,值得进一步的研究和探索,为计算机视觉领域做出贡献。
{"title":"Comparison and analysis of computer vision models based on images of catamount and canid","authors":"Feng Jiang, Yueyufei Ma, Langyue Wang","doi":"10.1117/12.2671468","DOIUrl":"https://doi.org/10.1117/12.2671468","url":null,"abstract":"Nowadays, target recognition, driverless, medical impact diagnosis, and other applications based on image recognition in life, scientific research, and work, rely mainly on a variety of large models with excellent performance, from the Convolutional Neural Network (CNN) at the beginning to the various variants of the classical model proposed now. In this paper, we will take the example of identifying catamount and canid datasets, comparing the efficiency and accuracy of CNN, Vision Transformer (ViT), and Swin Transformer laterally. We plan to run 25 epochs for each model and record the accuracy and time consumption separately. After the experiments we find that from the comparison of the epoch numbers and the real-time consumption, the CNN takes the least total time, followed by Swin Transformer. Also, ViT takes the least time to reach convergence, while Swin Transformer takes the most time. In terms of training accuracy, ViT has the highest training accuracy, followed by Swin Transformer, and CNN has the lowest training accuracy; the validation accuracy is similar to the training accuracy. ViT has the highest accuracy, but takes the longest time; conversely, CNN takes the shortest time and has the lowest accuracy. Swin Transformer, which seems a combination of CNN and ViT, is most complex but with ideal performance. In the future, ViT is indeed a promising model that deserves further research and exploration to contribute to the computer vision field.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127130243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Research on university network public opinion sentiment analysis based on BERT and Bi-LSTM 基于BERT和Bi-LSTM的高校网络舆情分析研究
Fangju Ran, Chen Xiong, Meng-yao Lu, Tianqing Yang
This paper proposes a method of emotion analysis based on BERT BiLSTM. Firstly, BERT is used to realize the word vectorization, and then Bilstm is constructed to extract semantic features for emotional analysis. In the experiment, the model designed in this paper is compared with the emotional dictionary, SVM, Word2vec LSTM, BERT TextCNN on the college online public opinion comment dataset, and the experiment proves that the accuracy of this model has been improved.
提出了一种基于BERT BiLSTM的情感分析方法。首先利用BERT实现词矢量化,然后构造Bilstm提取语义特征进行情感分析。在实验中,将本文设计的模型与情感词典、SVM、Word2vec LSTM、BERT TextCNN在高校在线舆情评论数据集上进行了对比,实验证明该模型的准确率得到了提高。
{"title":"Research on university network public opinion sentiment analysis based on BERT and Bi-LSTM","authors":"Fangju Ran, Chen Xiong, Meng-yao Lu, Tianqing Yang","doi":"10.1117/12.2671058","DOIUrl":"https://doi.org/10.1117/12.2671058","url":null,"abstract":"This paper proposes a method of emotion analysis based on BERT BiLSTM. Firstly, BERT is used to realize the word vectorization, and then Bilstm is constructed to extract semantic features for emotional analysis. In the experiment, the model designed in this paper is compared with the emotional dictionary, SVM, Word2vec LSTM, BERT TextCNN on the college online public opinion comment dataset, and the experiment proves that the accuracy of this model has been improved.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"199 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127301986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-cost fusion stereo matching algorithm based on guided filter aggregation 基于引导滤波聚合的多代价融合立体匹配算法
Jingwen Liu, Xuedong Zhang
Aiming at the low matching accuracy of existing local stereo matching algorithms in weak texture areas, a local stereo matching algorithm based on multi-matching cost fusion and guided filtering cost aggregation with adaptive parameters is proposed. First, use the gradient direction to improve the gradient cost, and calculate the matching cost by combining the gradient cost with the Census transform and color cost. Secondly, the cost is aggregated by the guided filtering of adaptive parameters; Finally, the final disparity map is obtained through disparity calculation and multi-step disparity refinement. The improved algorithm is tested on 15 training sets on the Middlebury3 platform, and the average false matching rates of bad4.0 in all areas and non-occluded areas are 19.9% and 13.2%, respectively, which is improved compared with AD-Census and other algorithms.
针对现有局部立体匹配算法在弱纹理区域匹配精度低的问题,提出了一种基于多匹配代价融合和自适应参数引导滤波代价聚合的局部立体匹配算法。首先,利用梯度方向改进梯度代价,并将梯度代价与Census变换和颜色代价相结合计算匹配代价。其次,通过自适应参数的引导滤波对代价进行聚合;最后,通过视差计算和多步视差细化得到最终的视差图。改进后的算法在Middlebury3平台上的15个训练集上进行了测试,所有区域和未遮挡区域的平均错误匹配率为bad4.0,分别为19.9%和13.2%,与AD-Census等算法相比有所提高。
{"title":"Multi-cost fusion stereo matching algorithm based on guided filter aggregation","authors":"Jingwen Liu, Xuedong Zhang","doi":"10.1117/12.2671218","DOIUrl":"https://doi.org/10.1117/12.2671218","url":null,"abstract":"Aiming at the low matching accuracy of existing local stereo matching algorithms in weak texture areas, a local stereo matching algorithm based on multi-matching cost fusion and guided filtering cost aggregation with adaptive parameters is proposed. First, use the gradient direction to improve the gradient cost, and calculate the matching cost by combining the gradient cost with the Census transform and color cost. Secondly, the cost is aggregated by the guided filtering of adaptive parameters; Finally, the final disparity map is obtained through disparity calculation and multi-step disparity refinement. The improved algorithm is tested on 15 training sets on the Middlebury3 platform, and the average false matching rates of bad4.0 in all areas and non-occluded areas are 19.9% and 13.2%, respectively, which is improved compared with AD-Census and other algorithms.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123727848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
New estimation for spectral radius of Hadamard product Hadamard积谱半径的新估计
Qin Zhong, Chunyan Zhao, Xin Zhou, Y. Wang, Ling Li
For the Hadamard product of the matrices with non-negative entries, we study the new upper bound for the spectral radius by applying the characteristic value containing the domain theorem. This estimating formula only involves the entries of two non-negative matrices. Hence, the upper bound is easy to calculate in practical examples. An example is considered to illustrate our results.
对于非负项矩阵的Hadamard积,应用包含定义域定理的特征值,研究了谱半径的新上界。这个估计公式只涉及两个非负矩阵的元素。因此,上界在实际算例中很容易计算。通过一个例子来说明我们的结果。
{"title":"New estimation for spectral radius of Hadamard product","authors":"Qin Zhong, Chunyan Zhao, Xin Zhou, Y. Wang, Ling Li","doi":"10.1117/12.2671104","DOIUrl":"https://doi.org/10.1117/12.2671104","url":null,"abstract":"For the Hadamard product of the matrices with non-negative entries, we study the new upper bound for the spectral radius by applying the characteristic value containing the domain theorem. This estimating formula only involves the entries of two non-negative matrices. Hence, the upper bound is easy to calculate in practical examples. An example is considered to illustrate our results.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122704438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
E-OrbF: a robust image feature matching algorithm E-OrbF:鲁棒图像特征匹配算法
Chang Liu, Huan Li
To improve the real-time performance and robustness of traditional feature matching algorithms, an improved image feature matching algorithm E-OrbF based on ORB and FREAK is proposed. In E-OrbF, the original FAST feature points in ORB algorithm are distributed unevenly and redundant. The strategy of subregion and local threshold is adopted to improve the uniform distribution and stability of feature points. Then simplify the sampling mode of FREAK algorithm and design a new feature descriptor. While improving the matching speed, the sampling point pairs are further filtered to improve the matching accuracy. Finally, combine RANSAC matching algorithm to eliminate mismatches and reduce the rate of mismatches. The experimental results show that the algorithm has good real-time performance, while under the conditions of perspective transformation, rotation scale, complex illumination and blur. Both of them can well complete feature detection and feature matching and improve the robustness of existing methods. The algorithm can be applied to the fusion of virtual and real scenes on mobile terminals, and the average visual frame rate reaches 30 FPS, meeting the real-time requirements.
为了提高传统特征匹配算法的实时性和鲁棒性,提出了一种基于ORB和FREAK的改进图像特征匹配算法E-OrbF。在E-OrbF中,ORB算法中原有的FAST特征点分布不均匀且冗余。采用子区域和局部阈值策略,提高特征点分布的均匀性和稳定性。然后简化了FREAK算法的采样方式,设计了新的特征描述符。在提高匹配速度的同时,进一步对采样点对进行滤波,提高匹配精度。最后结合RANSAC匹配算法消除错配,降低错配率。实验结果表明,该算法在透视变换、旋转尺度、复杂光照和模糊等条件下具有良好的实时性。两者都能很好地完成特征检测和特征匹配,提高了现有方法的鲁棒性。该算法可应用于移动端虚实场景融合,平均视觉帧率达到30 FPS,满足实时性要求。
{"title":"E-OrbF: a robust image feature matching algorithm","authors":"Chang Liu, Huan Li","doi":"10.1117/12.2671148","DOIUrl":"https://doi.org/10.1117/12.2671148","url":null,"abstract":"To improve the real-time performance and robustness of traditional feature matching algorithms, an improved image feature matching algorithm E-OrbF based on ORB and FREAK is proposed. In E-OrbF, the original FAST feature points in ORB algorithm are distributed unevenly and redundant. The strategy of subregion and local threshold is adopted to improve the uniform distribution and stability of feature points. Then simplify the sampling mode of FREAK algorithm and design a new feature descriptor. While improving the matching speed, the sampling point pairs are further filtered to improve the matching accuracy. Finally, combine RANSAC matching algorithm to eliminate mismatches and reduce the rate of mismatches. The experimental results show that the algorithm has good real-time performance, while under the conditions of perspective transformation, rotation scale, complex illumination and blur. Both of them can well complete feature detection and feature matching and improve the robustness of existing methods. The algorithm can be applied to the fusion of virtual and real scenes on mobile terminals, and the average visual frame rate reaches 30 FPS, meeting the real-time requirements.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131290594","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fact relation and keywords fusion abstractive summarization 事实关系与关键词的融合抽象概括
Shihao Tian, Long Zhang, Qiusheng Zheng
With the wide application of deep learning, the abstractive text summary has become an important research topic in natural language processing. The abstractive text summary has high flexibility and can generate words that have not appeared in the text. However, the generated summary model will have factual errors, which significantly affect the usability of the summary. Therefore, this paper proposes a text summary model based on fact relationships and keyword fusion. We extract the fact relation triplet in the input text and automatically extract the keywords in the text to assist in the generation of the abstract. The fusion of fact relations and keywords can effectively alleviate the problem of factual errors in the abstract. Many experiments show that compared with other baseline models, our model (FRKFS) improves the performance of summaries generated on the data sets CNN/Daily Mail and XSum and alleviates the problem of factual errors.
随着深度学习的广泛应用,抽象文本摘要已成为自然语言处理中的一个重要研究课题。抽象文本摘要具有很高的灵活性,可以生成文本中没有出现过的单词。然而,生成的摘要模型会有事实错误,这将严重影响摘要的可用性。为此,本文提出了一种基于事实关系和关键词融合的文本摘要模型。我们从输入文本中提取事实关系三元组,并自动提取文本中的关键字来辅助摘要的生成。事实关系与关键词的融合可以有效缓解摘要事实错误问题。许多实验表明,与其他基线模型相比,我们的模型(FRKFS)提高了在CNN/Daily Mail和XSum数据集上生成的摘要的性能,减轻了事实错误的问题。
{"title":"Fact relation and keywords fusion abstractive summarization","authors":"Shihao Tian, Long Zhang, Qiusheng Zheng","doi":"10.1117/12.2671188","DOIUrl":"https://doi.org/10.1117/12.2671188","url":null,"abstract":"With the wide application of deep learning, the abstractive text summary has become an important research topic in natural language processing. The abstractive text summary has high flexibility and can generate words that have not appeared in the text. However, the generated summary model will have factual errors, which significantly affect the usability of the summary. Therefore, this paper proposes a text summary model based on fact relationships and keyword fusion. We extract the fact relation triplet in the input text and automatically extract the keywords in the text to assist in the generation of the abstract. The fusion of fact relations and keywords can effectively alleviate the problem of factual errors in the abstract. Many experiments show that compared with other baseline models, our model (FRKFS) improves the performance of summaries generated on the data sets CNN/Daily Mail and XSum and alleviates the problem of factual errors.","PeriodicalId":227528,"journal":{"name":"International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-04-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126450009","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
International Conference on Artificial Intelligence and Computer Engineering (ICAICE 2022)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1