首页 > 最新文献

Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System最新文献

英文 中文
Fake Review Identification Method Based on Topic Model and Att-BiLSTM 基于主题模型和at - bilstm的虚假评论识别方法
Lei Shi, Suzhen Xie, Yongcai Tao, Lin Wei, Yufei Gao
The review rating system provides valuable information to potential users, but it also encourages the creation of profit-driven fake reviews. Fake reviews and comments not only drive consumers to buy low-quality products or services, but also erode consumers' long-term confidence in review rating platforms. At present, two main reasons for the low detection accuracy of fake comments in recent studies are: (1) lack of feature learning of emotional intensity of text; (2) the inaccuracy of the identification of topic words in comments. To solve the above problems, we propose a novel identification method based on topic model and Att-BiLSTM mechanism. The proposed method calculates text affective and subjective values using TextBlob, incorporating the topic feature to train the classifier for fake review recognition. Comparative experiments show that the model effect is better than other models.
评论评级系统为潜在用户提供了有价值的信息,但它也鼓励了以利润为导向的虚假评论的产生。虚假评论和评论不仅会促使消费者购买低质量的产品或服务,还会削弱消费者对评论评级平台的长期信心。目前,目前研究中假评论检测准确率低的主要原因有两个:(1)缺乏对文本情感强度的特征学习;(2)评论中主题词的识别不准确。为了解决上述问题,我们提出了一种基于主题模型和at - bilstm机制的识别方法。该方法利用TextBlob计算文本情感值和主观值,结合主题特征训练分类器进行虚假评论识别。对比实验表明,该模型的效果优于其他模型。
{"title":"Fake Review Identification Method Based on Topic Model and Att-BiLSTM","authors":"Lei Shi, Suzhen Xie, Yongcai Tao, Lin Wei, Yufei Gao","doi":"10.1145/3483845.3483881","DOIUrl":"https://doi.org/10.1145/3483845.3483881","url":null,"abstract":"The review rating system provides valuable information to potential users, but it also encourages the creation of profit-driven fake reviews. Fake reviews and comments not only drive consumers to buy low-quality products or services, but also erode consumers' long-term confidence in review rating platforms. At present, two main reasons for the low detection accuracy of fake comments in recent studies are: (1) lack of feature learning of emotional intensity of text; (2) the inaccuracy of the identification of topic words in comments. To solve the above problems, we propose a novel identification method based on topic model and Att-BiLSTM mechanism. The proposed method calculates text affective and subjective values using TextBlob, incorporating the topic feature to train the classifier for fake review recognition. Comparative experiments show that the model effect is better than other models.","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134477707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Application of AR in 3D model AR在3D模型中的应用
Liang Ma
In recent years, with the development and application of AR technology, more and more augmented reality applications have begun to appear in education, introduction, etc. to achieve display through specific 3D models, such as popularizing human body information through human skeleton models, and introducing cars' composition information through car models. As a brand-new interactive method, augmented reality-AR system can provide more detailed information, for the human by the direct-viewing feeling, and improve the efficiency of understanding information.
近年来,随着AR技术的发展和应用,越来越多的增强现实应用开始出现在教育、介绍等方面,通过特定的3D模型来实现展示,如通过人体骨骼模型来普及人体信息,通过汽车模型来介绍汽车的成分信息等。增强现实-增强现实系统作为一种全新的交互方式,可以为人类提供更详细的信息,通过直观的感受,提高对信息的理解效率。
{"title":"Application of AR in 3D model","authors":"Liang Ma","doi":"10.1145/3483845.3483891","DOIUrl":"https://doi.org/10.1145/3483845.3483891","url":null,"abstract":"In recent years, with the development and application of AR technology, more and more augmented reality applications have begun to appear in education, introduction, etc. to achieve display through specific 3D models, such as popularizing human body information through human skeleton models, and introducing cars' composition information through car models. As a brand-new interactive method, augmented reality-AR system can provide more detailed information, for the human by the direct-viewing feeling, and improve the efficiency of understanding information.","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131709280","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Parallel Accelerated Algorithm Optimization for Speckle Matching in Deformation Measurement Based on Stereo Vision 基于立体视觉的形变测量中散斑匹配并行加速算法优化
Yunhe Liu, Guiyang Zhang, Lili Wang, Jing Wang, Zijian Zhu
This paper is concerned with the efficiency of speckle match in vision deformation measurement, upon which the CUDA programming architecture, combined with the Visual Studio platform and Mex script files is utilized to implement parallel operations. With the aid of compiling the GPU parallel mode of the CUDA source program through NVCC, the scheme of speckle matching parallel computing are given, which is crucial to improve the real-time performance of vision-based deformation measurement. Consequently, the method in this paper completes the efficient calculation of match of the speckle image sub-regions in the three-dimensional deformation measurement. The proposed strategy solves the obstacle problem when the Mex script and different programming languages interact, and is not restricted by overloaded functions, so that the overall computing performance of the deformation measurement program reaches a better state. Lastly, the experimental results show that the speckle matching has achieved a calculation speedup ratio of 20.39 times.
本文研究了视觉变形测量中散斑匹配的效率问题,在此基础上利用CUDA编程架构,结合Visual Studio平台和Mex脚本文件实现并行运算。通过NVCC编译CUDA源程序的GPU并行模式,给出了散斑匹配并行计算方案,这对提高基于视觉的变形测量的实时性至关重要。因此,本文方法完成了三维变形测量中散斑图像子区域匹配的高效计算。该策略解决了Mex脚本与不同编程语言交互时的障碍问题,且不受重载函数的限制,使变形测量程序的整体计算性能达到较好的状态。最后,实验结果表明,散斑匹配的计算加速比达到了20.39倍。
{"title":"Parallel Accelerated Algorithm Optimization for Speckle Matching in Deformation Measurement Based on Stereo Vision","authors":"Yunhe Liu, Guiyang Zhang, Lili Wang, Jing Wang, Zijian Zhu","doi":"10.1145/3483845.3483889","DOIUrl":"https://doi.org/10.1145/3483845.3483889","url":null,"abstract":"This paper is concerned with the efficiency of speckle match in vision deformation measurement, upon which the CUDA programming architecture, combined with the Visual Studio platform and Mex script files is utilized to implement parallel operations. With the aid of compiling the GPU parallel mode of the CUDA source program through NVCC, the scheme of speckle matching parallel computing are given, which is crucial to improve the real-time performance of vision-based deformation measurement. Consequently, the method in this paper completes the efficient calculation of match of the speckle image sub-regions in the three-dimensional deformation measurement. The proposed strategy solves the obstacle problem when the Mex script and different programming languages interact, and is not restricted by overloaded functions, so that the overall computing performance of the deformation measurement program reaches a better state. Lastly, the experimental results show that the speckle matching has achieved a calculation speedup ratio of 20.39 times.","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130299710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploratory Analysis on Topic Modelling for Video Subtitles 视频字幕主题建模的探索性分析
Atmik Ajoy, Chethan U Mahindrakar, H. Mamatha
In this paper, we explore different models available to perform topic modelling on subtitles files. Subtitle files are sourced from movies and represent the dialogue being spoken. Applying this to topic modelling would mean trying to obtain the topics regarding the video from only the subtitles. Our novel idea is to test whether it would be feasible to use topic modelling on subtitles to get topics of a movie. While topic modelling as an idea has been used previously in bio-informatics,patent indexing and much more, has not seen any application in this sphere. We extensively search for datasets, preprocess the subtitles files and try Latent Dirichlet Allocation, Hierarchical Dirichlet Processes and Latent Semantic Indexing methods of topic modelling on these documents. These are the top three prominent topic modelling models that are used today. Our results entail what model would work best for subtitle files
在本文中,我们探索了不同的模型来对字幕文件进行主题建模。字幕文件来自电影,代表正在说话的对话。将此应用于主题建模将意味着试图仅从字幕中获取有关视频的主题。我们的新颖想法是测试在字幕上使用主题建模来获取电影主题是否可行。虽然主题建模作为一种思想已经在生物信息学、专利索引等领域得到了应用,但在这一领域还没有任何应用。我们广泛搜索数据集,对字幕文件进行预处理,并尝试对这些文档进行潜在狄利克雷分配、层次狄利克雷过程和潜在语义索引等主题建模方法。这是目前使用的三个最突出的主题建模模型。我们的结果确定了哪种模型最适合字幕文件
{"title":"Exploratory Analysis on Topic Modelling for Video Subtitles","authors":"Atmik Ajoy, Chethan U Mahindrakar, H. Mamatha","doi":"10.1145/3483845.3483878","DOIUrl":"https://doi.org/10.1145/3483845.3483878","url":null,"abstract":"In this paper, we explore different models available to perform topic modelling on subtitles files. Subtitle files are sourced from movies and represent the dialogue being spoken. Applying this to topic modelling would mean trying to obtain the topics regarding the video from only the subtitles. Our novel idea is to test whether it would be feasible to use topic modelling on subtitles to get topics of a movie. While topic modelling as an idea has been used previously in bio-informatics,patent indexing and much more, has not seen any application in this sphere. We extensively search for datasets, preprocess the subtitles files and try Latent Dirichlet Allocation, Hierarchical Dirichlet Processes and Latent Semantic Indexing methods of topic modelling on these documents. These are the top three prominent topic modelling models that are used today. Our results entail what model would work best for subtitle files","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"194 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122575724","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improved YOLOv5 network-based object detection for anti-intrusion of gantry crane 基于改进YOLOv5网络的龙门起重机防入侵目标检测
Hongchao Niu, Xiao-Bing Hu, Hang Li
In response to the current lack of intelligence and security research on outdoor gantry cranes, the method based on the improved you-only-look-once (YOLO)v5 network for intelligent anti-intrusion detection is proposed. First an overall detection scheme is proposed. Then the following improvement tricks are made to the YOLOv5 network to achieve the highest possible detection accuracy while ensuring speed: incorporate multi-layer receptive fields and fine-grained modules into the backbone network to improve the performance of features; use dilated convolution to replace the pooling operation in the SPP module to reduce the loss of network information; further enrich the fusion of non-adjacent deep and shallow features in the network by using cross-layer connections; then use the K-means algorithm to cluster the target size to improve the positioning accuracy of the model; Finally, the non-maximum suppression algorithm is optimized by the weighting algorithm to effectively alleviate the inaccurate positioning of the YOLO series of bounding boxes. By combining multiple tricks, the improved YOLOv5s model can achieve a better balance between effectiveness (75.81% mAP) and efficiency (83 FPS) in anti-intrusion detection. At the same time, compared with the original YOLOv5s network on the VOC data set, the mAP value of the improved YOLOv5s is increased by 7.05%.
针对目前户外龙门起重机智能与安全研究的不足,提出了一种基于改进you-only-look-once (YOLO)v5网络的智能防入侵检测方法。首先提出了一种整体检测方案。然后对YOLOv5网络进行以下改进技巧,以在保证速度的同时达到尽可能高的检测精度:在骨干网络中加入多层接受域和细粒度模块,以提高特征的性能;采用扩展卷积代替SPP模块中的池化操作,减少网络信息的丢失;利用跨层连接进一步丰富了网络中非相邻深浅特征的融合;然后利用K-means算法对目标大小进行聚类,提高模型的定位精度;最后,通过加权算法对非最大抑制算法进行优化,有效缓解了YOLO系列边界框定位不准确的问题。通过结合多种技巧,改进的YOLOv5s模型可以更好地平衡入侵检测的有效性(75.81% mAP)和效率(83 FPS)。同时,在VOC数据集上,改进后的YOLOv5s网络的mAP值比原始YOLOv5s网络的mAP值提高了7.05%。
{"title":"Improved YOLOv5 network-based object detection for anti-intrusion of gantry crane","authors":"Hongchao Niu, Xiao-Bing Hu, Hang Li","doi":"10.1145/3483845.3483871","DOIUrl":"https://doi.org/10.1145/3483845.3483871","url":null,"abstract":"In response to the current lack of intelligence and security research on outdoor gantry cranes, the method based on the improved you-only-look-once (YOLO)v5 network for intelligent anti-intrusion detection is proposed. First an overall detection scheme is proposed. Then the following improvement tricks are made to the YOLOv5 network to achieve the highest possible detection accuracy while ensuring speed: incorporate multi-layer receptive fields and fine-grained modules into the backbone network to improve the performance of features; use dilated convolution to replace the pooling operation in the SPP module to reduce the loss of network information; further enrich the fusion of non-adjacent deep and shallow features in the network by using cross-layer connections; then use the K-means algorithm to cluster the target size to improve the positioning accuracy of the model; Finally, the non-maximum suppression algorithm is optimized by the weighting algorithm to effectively alleviate the inaccurate positioning of the YOLO series of bounding boxes. By combining multiple tricks, the improved YOLOv5s model can achieve a better balance between effectiveness (75.81% mAP) and efficiency (83 FPS) in anti-intrusion detection. At the same time, compared with the original YOLOv5s network on the VOC data set, the mAP value of the improved YOLOv5s is increased by 7.05%.","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121607307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
An Efficient Addressing Scheme for Flexible IP Address 灵活IP地址的高效寻址方案
Shi-Hai Liu, Wanming Luo, Xu Zhou, YiHao Jia, Zhe Chen, Sheng Jiang
Along with the popularization and adoption of IP in various emerging scenarios, challenges also arise with the ossified address structures. The reason is that conventional IP address is designed with fixed length and lacking extensibility, while the demand for IP varies greatly in different scenarios. Flexible IP (FlexIP), as a variable-length IP address, proactively makes address structure flexible enough to adapt to various network cases and solves the problem of low transmission efficiency faced by current IP addresses. However, due to the variable length of FlexIP, the conventional routing addressing scheme is not suitable for it. In this paper, we propose a new Bloom filter addressing scheme suitable for FlexIP address. We use controllable prefix extension to limit the prefix distribution of FlexIP, and use one-hashing to improve the computational overhead of the Bloom filter. Simulations show that the addressing scheme we proposed is more suitable for FlexIP than other schemes, and has better query efficiency.
随着IP在各种新兴场景中的普及和采用,地址结构的僵化也带来了挑战。原因是传统的IP地址设计固定长度,缺乏可扩展性,而不同场景对IP的需求差异很大。FlexIP (Flexible IP)是一种可变长度的IP地址,它主动使地址结构具有足够的灵活性,以适应各种网络情况,解决了当前IP地址传输效率低的问题。然而,由于FlexIP的长度可变,传统的路由寻址方案不适合它。在本文中,我们提出了一种新的适用于FlexIP地址的Bloom滤波器寻址方案。我们使用可控前缀扩展来限制FlexIP的前缀分布,并使用单哈希来改善Bloom过滤器的计算开销。仿真结果表明,该寻址方案比其他寻址方案更适合FlexIP,具有更好的查询效率。
{"title":"An Efficient Addressing Scheme for Flexible IP Address","authors":"Shi-Hai Liu, Wanming Luo, Xu Zhou, YiHao Jia, Zhe Chen, Sheng Jiang","doi":"10.1145/3483845.3483865","DOIUrl":"https://doi.org/10.1145/3483845.3483865","url":null,"abstract":"Along with the popularization and adoption of IP in various emerging scenarios, challenges also arise with the ossified address structures. The reason is that conventional IP address is designed with fixed length and lacking extensibility, while the demand for IP varies greatly in different scenarios. Flexible IP (FlexIP), as a variable-length IP address, proactively makes address structure flexible enough to adapt to various network cases and solves the problem of low transmission efficiency faced by current IP addresses. However, due to the variable length of FlexIP, the conventional routing addressing scheme is not suitable for it. In this paper, we propose a new Bloom filter addressing scheme suitable for FlexIP address. We use controllable prefix extension to limit the prefix distribution of FlexIP, and use one-hashing to improve the computational overhead of the Bloom filter. Simulations show that the addressing scheme we proposed is more suitable for FlexIP than other schemes, and has better query efficiency.","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130502894","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Synchronized Multi-Helical Computed Tomography 同步多螺旋计算机断层扫描
Changsheng Zhang, Guogang Zhu, Jian Fu
Limited by the field of view (FOV), most existed X-ray industrial computed tomography (ICT) techniques require multi scans for stitching projections when detecting long objects, which significantly increases the scanning time. In addition, these techniques usually adopt the one-by-one scanning mode that further reduces the scanning efficiency. Therefore, this paper proposes a synchronized multi-helical computed tomography. It allows multi objects to be helical scanned simultaneously without signal crosstalk, while it further improves the detecting efficiency. Besides, the reconstruction method suitable for the synchronized multi-helical CT is reported. This method utilizes projection segmentation and helical projection calibration to convert multi-object helical projections into single-object projections. The generated single-object projection can be then reconstructed by conventional algorithms, e.g. the filtered back projection (FBP). This work can improve the efficiency of CT scanning and will promote the applications of CT in large-scale long object detection.
现有的x射线工业计算机断层扫描(ICT)技术在检测长物体时,由于视场(FOV)的限制,需要对拼接投影进行多次扫描,这大大增加了扫描时间。此外,这些技术通常采用一对一扫描方式,进一步降低了扫描效率。因此,本文提出了一种同步的多螺旋计算机断层扫描方法。它可以同时对多个目标进行螺旋扫描,没有信号串扰,进一步提高了检测效率。此外,还报道了适用于同步多螺旋CT的重建方法。该方法利用投影分割和螺旋投影标定将多目标螺旋投影转化为单目标投影。生成的单目标投影可以通过传统算法重建,例如滤波后的反投影(FBP)。这项工作可以提高CT扫描的效率,促进CT在大规模长目标检测中的应用。
{"title":"Synchronized Multi-Helical Computed Tomography","authors":"Changsheng Zhang, Guogang Zhu, Jian Fu","doi":"10.1145/3483845.3483883","DOIUrl":"https://doi.org/10.1145/3483845.3483883","url":null,"abstract":"Limited by the field of view (FOV), most existed X-ray industrial computed tomography (ICT) techniques require multi scans for stitching projections when detecting long objects, which significantly increases the scanning time. In addition, these techniques usually adopt the one-by-one scanning mode that further reduces the scanning efficiency. Therefore, this paper proposes a synchronized multi-helical computed tomography. It allows multi objects to be helical scanned simultaneously without signal crosstalk, while it further improves the detecting efficiency. Besides, the reconstruction method suitable for the synchronized multi-helical CT is reported. This method utilizes projection segmentation and helical projection calibration to convert multi-object helical projections into single-object projections. The generated single-object projection can be then reconstructed by conventional algorithms, e.g. the filtered back projection (FBP). This work can improve the efficiency of CT scanning and will promote the applications of CT in large-scale long object detection.","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133573043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effect of regularity on learning in GANs 规则性对gan学习的影响
Niladri Shekhar Dutt, S. Patel
Generative Adversarial Networks (GANs) are algorithmic architectures that use two neural networks, pitting one against the opposite (thus the “adversarial”) so as to come up with new, synthetic instances of data that can pass for real data. GANs have been highly successful on datasets like MNIST, SVHN, CelebA, etc but training a GAN on large scale datasets like ImageNet is a challenging problem because they are deemed as not very regular. In this paper, we perform empirical experiments using parameterized synthetic datasets to probe how regularity of a dataset affects learning in GANs. We emperically show that regular datasets are easier to model for GANs because of their stable training process.
生成对抗网络(GANs)是一种算法架构,它使用两个神经网络,让一个神经网络对抗另一个神经网络(因此称为“对抗”),从而产生新的、合成的数据实例,这些数据实例可以被当作真实数据。GAN在MNIST、SVHN、CelebA等数据集上非常成功,但在像ImageNet这样的大规模数据集上训练GAN是一个具有挑战性的问题,因为它们被认为不是很有规律。在本文中,我们使用参数化合成数据集进行经验实验,以探索数据集的规律性如何影响gan中的学习。我们的经验表明,由于正则数据集的训练过程稳定,因此更容易对gan进行建模。
{"title":"Effect of regularity on learning in GANs","authors":"Niladri Shekhar Dutt, S. Patel","doi":"10.1145/3483845.3483874","DOIUrl":"https://doi.org/10.1145/3483845.3483874","url":null,"abstract":"Generative Adversarial Networks (GANs) are algorithmic architectures that use two neural networks, pitting one against the opposite (thus the “adversarial”) so as to come up with new, synthetic instances of data that can pass for real data. GANs have been highly successful on datasets like MNIST, SVHN, CelebA, etc but training a GAN on large scale datasets like ImageNet is a challenging problem because they are deemed as not very regular. In this paper, we perform empirical experiments using parameterized synthetic datasets to probe how regularity of a dataset affects learning in GANs. We emperically show that regular datasets are easier to model for GANs because of their stable training process.","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133378821","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Learning A Linear Classifier by Transforming Feature Vectors for Few-shot Image Classification 基于变换特征向量的线性分类器的学习
Wanrong Huang, Yaqing Hu, Shuofeng Hu, Jingde Liu
Deep neural networks have achieved remarkable results in large-scale data domain. However, they have not performed well on few-shot image classification tasks. Here we propose a new meta-learning approach composed of an embedding network and a linear classifier learner. During the training phase, our approach (called Transformation Network) learns to learn a classifier by transforming the feature vectors produced by the embedding module. Once trained, a Transformation Network is able to classify images of new classes by the learned classifier. The ability of learning a discriminatively trained classifier could make our architecture adapt fast to new examples from unseen classes. We further describe implementation details upon the architecture convolutional networks and linear transformation operations. We demonstrate that our approach achieves improved performance on few-shot image classification tasks on two benchmarks and a self-made dataset.
深度神经网络在大规模数据领域取得了显著的成果。然而,它们在少量图像分类任务中表现不佳。本文提出了一种由嵌入网络和线性分类器学习器组成的元学习方法。在训练阶段,我们的方法(称为转换网络)通过转换嵌入模块产生的特征向量来学习学习分类器。经过训练后,转换网络能够通过学习到的分类器对新类别的图像进行分类。学习判别训练分类器的能力可以使我们的体系结构快速适应来自未知类的新示例。我们进一步描述了架构卷积网络和线性变换操作的实现细节。我们在两个基准测试和一个自制数据集上证明了我们的方法在少量图像分类任务上取得了更好的性能。
{"title":"Learning A Linear Classifier by Transforming Feature Vectors for Few-shot Image Classification","authors":"Wanrong Huang, Yaqing Hu, Shuofeng Hu, Jingde Liu","doi":"10.1145/3483845.3483873","DOIUrl":"https://doi.org/10.1145/3483845.3483873","url":null,"abstract":"Deep neural networks have achieved remarkable results in large-scale data domain. However, they have not performed well on few-shot image classification tasks. Here we propose a new meta-learning approach composed of an embedding network and a linear classifier learner. During the training phase, our approach (called Transformation Network) learns to learn a classifier by transforming the feature vectors produced by the embedding module. Once trained, a Transformation Network is able to classify images of new classes by the learned classifier. The ability of learning a discriminatively trained classifier could make our architecture adapt fast to new examples from unseen classes. We further describe implementation details upon the architecture convolutional networks and linear transformation operations. We demonstrate that our approach achieves improved performance on few-shot image classification tasks on two benchmarks and a self-made dataset.","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132440671","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bronchial Light Microscopy Image Segmentation Based on Boundary Attention 基于边界关注的支气管光学显微镜图像分割
Kuncheng Li, Zhexin Li, Yicheng Liu, Qinzhi Fang, Bangwangke Tang, Liping Huang, Xinyu Xiong
∗The identification of bronchus is of great significance in assisting the diagnosis of lung diseases. However, identifying the bronchus from tissue light microscopy images is a heavily repetitive task that requires a lot of time and effort. Most of the mainstream segmentation methods pay attention to the overall accuracy of the region, without special consideration for the boundaries. However, bronchi often have flexible shapes, which poses a challenge for accurate segmentation, especially for details at the edges. Therefore, this paper proposes a boundary-attention based bronchi segmentation network. This network is a “predict and refine” architecture. Specifically, a coarse segmentation result is first generated by the prediction network, and then the edge segmentation quality is improved by the refinement network. In addition, by specially designed hybrid loss, our network can focus on patch-level contextual information as well as pixel-level accuracy. At the same time, the global attention module and the local attention module enable our network to extract both multiscale features and focus on error-prone regions. Through our network, not only fine segmentation results can be achieved, but also superior performance at the bronchial boundary. Experiments on the BronSeg dataset show that our method outperforms mainstream methods in all metrics, especially in mIOU, which reaches 88.41%. ∗Corresponding author. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from permissions@acm.org. CCRIS’21, August 20–22, 2021, Qingdao, China © 2021 Association for Computing Machinery. ACM ISBN 978-1-4503-9045-3/21/08. . . $15.00 https://doi.org/10.1145/3483845.3483890 CCS CONCEPTS • Artificial intelligence; • Computer vision; • Image segmentation;
*支气管的鉴别对于协助肺部疾病的诊断有重要的意义。然而,从组织光学显微镜图像中识别支气管是一项非常重复的任务,需要大量的时间和精力。主流的分割方法大多关注区域的整体精度,而没有特别考虑区域的边界。然而,支气管通常具有灵活的形状,这对准确分割提出了挑战,特别是对边缘的细节。为此,本文提出了一种基于边界注意的支气管分割网络。这个网络是一个“预测和改进”的架构。具体而言,首先由预测网络生成粗分割结果,然后通过细化网络提高边缘分割质量。此外,通过特殊设计的混合损失,我们的网络可以专注于补丁级上下文信息和像素级精度。同时,全局关注模块和局部关注模块使我们的网络既可以提取多尺度特征,又可以关注容易出错的区域。通过我们的网络,不仅可以获得良好的分割效果,而且在支气管边界处表现优异。在BronSeg数据集上的实验表明,我们的方法在所有指标上都优于主流方法,特别是在mIOU上达到了88.41%。∗通讯作者。允许免费制作本作品的全部或部分数字或硬拷贝供个人或课堂使用,前提是副本不是为了盈利或商业利益而制作或分发的,并且副本在第一页上带有本通知和完整的引用。本作品组件的版权归ACM以外的其他人所有,必须得到尊重。允许有信用的摘要。以其他方式复制或重新发布,在服务器上发布或重新分发到列表,需要事先获得特定许可和/或付费。从permissions@acm.org请求权限。CCRIS ' 21, 2021年8月20-22日,中国青岛©2021计算机械协会。Acm isbn 978-1-4503-9045-3/21/08…$15.00 https://doi.org/10.1145/3483845.3483890 CCS CONCEPTS•人工智能;•计算机视觉;•图像分割;
{"title":"Bronchial Light Microscopy Image Segmentation Based on Boundary Attention","authors":"Kuncheng Li, Zhexin Li, Yicheng Liu, Qinzhi Fang, Bangwangke Tang, Liping Huang, Xinyu Xiong","doi":"10.1145/3483845.3483890","DOIUrl":"https://doi.org/10.1145/3483845.3483890","url":null,"abstract":"∗The identification of bronchus is of great significance in assisting the diagnosis of lung diseases. However, identifying the bronchus from tissue light microscopy images is a heavily repetitive task that requires a lot of time and effort. Most of the mainstream segmentation methods pay attention to the overall accuracy of the region, without special consideration for the boundaries. However, bronchi often have flexible shapes, which poses a challenge for accurate segmentation, especially for details at the edges. Therefore, this paper proposes a boundary-attention based bronchi segmentation network. This network is a “predict and refine” architecture. Specifically, a coarse segmentation result is first generated by the prediction network, and then the edge segmentation quality is improved by the refinement network. In addition, by specially designed hybrid loss, our network can focus on patch-level contextual information as well as pixel-level accuracy. At the same time, the global attention module and the local attention module enable our network to extract both multiscale features and focus on error-prone regions. Through our network, not only fine segmentation results can be achieved, but also superior performance at the bronchial boundary. Experiments on the BronSeg dataset show that our method outperforms mainstream methods in all metrics, especially in mIOU, which reaches 88.41%. ∗Corresponding author. Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from permissions@acm.org. CCRIS’21, August 20–22, 2021, Qingdao, China © 2021 Association for Computing Machinery. ACM ISBN 978-1-4503-9045-3/21/08. . . $15.00 https://doi.org/10.1145/3483845.3483890 CCS CONCEPTS • Artificial intelligence; • Computer vision; • Image segmentation;","PeriodicalId":134636,"journal":{"name":"Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117300334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Proceedings of the 2021 2nd International Conference on Control, Robotics and Intelligent System
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1