首页 > 最新文献

2020 2nd Symposium on Signal Processing Systems最新文献

英文 中文
An Improved Chinese Named Entity Recognition Method with TB-LSTM-CRF 一种改进的TB-LSTM-CRF中文命名实体识别方法
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421534
Jiazheng Li, Tao Wang, Weiwen Zhang
Owing to the lack of natural delimiters, Chinese named entity recognition (NER) is more challenging than it in English. While Chinese word segmentation (CWS) is generally regarded as key and open problem for Chinese NER, its accuracy is critical for the downstream models trainings and it also often suffers from out-of-vocabulary (OOV). In this paper, we propose an improved Chinese NER model called TB-LSTM-CRF, which introduces a Transformer Block on top of LSTM-CRF. The proposed model with Transformer Block exploits the self-attention mechanism to capture the information from adjacent characters and sentence contexts. It is more practical with using small-size character embeddings. Compared with the baseline using LSTM-CRF, experiment results show our method TB-LSTM-CRF is competitive without the need of any external resources, for instance other dictionaries.
由于缺乏自然分隔符,中文命名实体识别比英文命名实体识别更具挑战性。摘要中文分词(CWS)一直被认为是中文NER的关键和开放性问题,其准确性对下游模型的训练至关重要,并且经常出现词汇外(OOV)的问题。在本文中,我们提出了一种改进的中国NER模型TB-LSTM-CRF,该模型在LSTM-CRF之上引入了一个变压器块。该模型利用自注意机制从相邻的字符和句子上下文中捕获信息。使用小尺寸字符嵌入更实用。实验结果表明,与LSTM-CRF方法相比,TB-LSTM-CRF方法在不需要任何外部资源(如其他字典)的情况下具有一定的竞争力。
{"title":"An Improved Chinese Named Entity Recognition Method with TB-LSTM-CRF","authors":"Jiazheng Li, Tao Wang, Weiwen Zhang","doi":"10.1145/3421515.3421534","DOIUrl":"https://doi.org/10.1145/3421515.3421534","url":null,"abstract":"Owing to the lack of natural delimiters, Chinese named entity recognition (NER) is more challenging than it in English. While Chinese word segmentation (CWS) is generally regarded as key and open problem for Chinese NER, its accuracy is critical for the downstream models trainings and it also often suffers from out-of-vocabulary (OOV). In this paper, we propose an improved Chinese NER model called TB-LSTM-CRF, which introduces a Transformer Block on top of LSTM-CRF. The proposed model with Transformer Block exploits the self-attention mechanism to capture the information from adjacent characters and sentence contexts. It is more practical with using small-size character embeddings. Compared with the baseline using LSTM-CRF, experiment results show our method TB-LSTM-CRF is competitive without the need of any external resources, for instance other dictionaries.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126193440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Automatic Language Identification using Suprasegmental Feature and Supervised Topic Model 基于超切分特征和监督主题模型的语言自动识别
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421521
Linjia Sun
Language identification is quite challenging when it comes to discriminating between closely related dialects of the same language. The fundamental issue is to explore the discriminative cue and effective representation. In this paper, the multi-dimensional language cues are used to distinguish languages, which includes the phonotactic and prosodic information and can be found in the unsupervised setting. Moreover, a novel supervised topic model is proposed to represent and learn the difference of languages. We built the system of language identification and reported the test results on the NIST LRE07 datasets and the Chinese dialect spoken corpus. Compared with other state-of-the-art methods, the experiment results show that the proposed method provides competitive performance and helps to capture robust discriminative information for short duration language identification.
当涉及到在同一种语言的密切相关的方言之间进行区分时,语言识别是相当具有挑战性的。本研究的根本问题是探讨区别线索和有效表征。本文使用多维语言线索来区分语言,多维语言线索包括语音和韵律信息,这些信息可以在无监督环境中找到。此外,还提出了一种新的监督主题模型来表示和学习语言之间的差异。我们构建了语言识别系统,并在NIST LRE07数据集和汉语方言口语语料库上报告了测试结果。实验结果表明,该方法在短时间语言识别中具有较强的鲁棒性和较强的识别能力。
{"title":"Automatic Language Identification using Suprasegmental Feature and Supervised Topic Model","authors":"Linjia Sun","doi":"10.1145/3421515.3421521","DOIUrl":"https://doi.org/10.1145/3421515.3421521","url":null,"abstract":"Language identification is quite challenging when it comes to discriminating between closely related dialects of the same language. The fundamental issue is to explore the discriminative cue and effective representation. In this paper, the multi-dimensional language cues are used to distinguish languages, which includes the phonotactic and prosodic information and can be found in the unsupervised setting. Moreover, a novel supervised topic model is proposed to represent and learn the difference of languages. We built the system of language identification and reported the test results on the NIST LRE07 datasets and the Chinese dialect spoken corpus. Compared with other state-of-the-art methods, the experiment results show that the proposed method provides competitive performance and helps to capture robust discriminative information for short duration language identification.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123711015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Text Sentiment Analysis based on Parallel TCN Model and Attention Model 基于并行TCN模型和注意力模型的文本情感分析
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421524
Dong Cao, Yujie Huang, Yunbin Fu
Aiming at the problem that the traditional single convolutional neural network cannot completely extract comprehensive text features, this paper proposes a text sentiment classification based on the parallel TCN model of attention mechanism. First, obtain the comprehensive text features with the help of parallel Temporal Convolutional Network (TCN). Secondly, in the feature fusion layer, the features obtained by the parallel TCN are fused. Finally, it combines the attention mechanism to extract important feature information and improve the optimized text sentiment classification effect. And conducted multiple sets of comparative experiments on the two sets of Chinese data sets, the accuracy of the model in this paper reached 92.06% and 92.71%. Proved that the proposed model is better than the traditional single convolutional neural network.
针对传统单卷积神经网络不能完全提取文本综合特征的问题,提出了一种基于注意力机制的并行TCN模型的文本情感分类方法。首先,利用并行时间卷积网络(TCN)获取综合文本特征;其次,在特征融合层,对并行TCN得到的特征进行融合;最后结合注意机制提取重要特征信息,提高优化后的文本情感分类效果。并对两组中文数据集进行了多组对比实验,本文模型的准确率分别达到了92.06%和92.71%。证明了该模型优于传统的单卷积神经网络。
{"title":"Text Sentiment Analysis based on Parallel TCN Model and Attention Model","authors":"Dong Cao, Yujie Huang, Yunbin Fu","doi":"10.1145/3421515.3421524","DOIUrl":"https://doi.org/10.1145/3421515.3421524","url":null,"abstract":"Aiming at the problem that the traditional single convolutional neural network cannot completely extract comprehensive text features, this paper proposes a text sentiment classification based on the parallel TCN model of attention mechanism. First, obtain the comprehensive text features with the help of parallel Temporal Convolutional Network (TCN). Secondly, in the feature fusion layer, the features obtained by the parallel TCN are fused. Finally, it combines the attention mechanism to extract important feature information and improve the optimized text sentiment classification effect. And conducted multiple sets of comparative experiments on the two sets of Chinese data sets, the accuracy of the model in this paper reached 92.06% and 92.71%. Proved that the proposed model is better than the traditional single convolutional neural network.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129862917","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Pedestrian Re-identification Method Based on Multi-frame Fusion Part-based Convolutional Baseline Network 基于多帧融合部分卷积基线网络的行人再识别方法
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421533
Yuxiang Peng, Guoheng Huang, Tao Peng, Lianglun Cheng, Hui-Shi Wu
In recent years, with the increasingly perfect monitoring system, how to make full use of the existing monitoring system to do security work has become a concern in the security field. Face recognition can be used in the security field, but it is difficult to play a role in the surveillance field because it usually requires the cooperation of pedestrians. Therefore, the pedestrian recognition technology without the cooperation of pedestrians has been widely concerned. In this paper, in order to realize a given sequence of monitoring pedestrian images and retrieve pedestrian images across devices, we proposed a new method to realize high- precision pedestrian recognition. First, because surveillance video is a series of pedestrian sequences, we proposed a Crossover Filtering Module (CFM) to screen video sequences for key frames. Then, we propose a network named Multi-frame Fusion Part- based Convolutional Baseline (MFPCB) to extract the features of screened keyframes. Finally, we use the cosine distance to measure the features and find the pedestrian image across the device. This paper mainly studies feature comparison and extraction, which can solve the problems of pedestrian occlusion and location under different cameras. Experiment confirms that MFPCB allows pedestrian recognition to gain another round of performance boost. For instance, on the Mars dataset, we achieve 77.3% mAP and 88.6% rank-1 accuracy, surpassing the state of the art by a large margin.
近年来,随着监控系统的日益完善,如何充分利用现有的监控系统做好安防工作成为安防领域关注的问题。人脸识别可以应用于安防领域,但由于通常需要行人的配合,很难在监控领域发挥作用。因此,不需要行人配合的行人识别技术受到了广泛关注。为了实现给定序列的行人图像监控和跨设备检索行人图像,提出了一种实现高精度行人识别的新方法。首先,由于监控视频是一系列行人序列,我们提出了一个交叉滤波模块(CFM)来筛选视频序列的关键帧。然后,我们提出了一种基于多帧融合部分的卷积基线(MFPCB)网络来提取筛选出的关键帧的特征。最后,我们使用余弦距离来测量特征并找到跨设备的行人图像。本文主要研究特征比较与提取,解决不同摄像头下行人遮挡与定位问题。实验证实,MFPCB可以使行人识别获得另一轮性能提升。例如,在火星数据集上,我们实现了77.3%的mAP和88.6%的rank-1精度,大大超过了目前的水平。
{"title":"A Pedestrian Re-identification Method Based on Multi-frame Fusion Part-based Convolutional Baseline Network","authors":"Yuxiang Peng, Guoheng Huang, Tao Peng, Lianglun Cheng, Hui-Shi Wu","doi":"10.1145/3421515.3421533","DOIUrl":"https://doi.org/10.1145/3421515.3421533","url":null,"abstract":"In recent years, with the increasingly perfect monitoring system, how to make full use of the existing monitoring system to do security work has become a concern in the security field. Face recognition can be used in the security field, but it is difficult to play a role in the surveillance field because it usually requires the cooperation of pedestrians. Therefore, the pedestrian recognition technology without the cooperation of pedestrians has been widely concerned. In this paper, in order to realize a given sequence of monitoring pedestrian images and retrieve pedestrian images across devices, we proposed a new method to realize high- precision pedestrian recognition. First, because surveillance video is a series of pedestrian sequences, we proposed a Crossover Filtering Module (CFM) to screen video sequences for key frames. Then, we propose a network named Multi-frame Fusion Part- based Convolutional Baseline (MFPCB) to extract the features of screened keyframes. Finally, we use the cosine distance to measure the features and find the pedestrian image across the device. This paper mainly studies feature comparison and extraction, which can solve the problems of pedestrian occlusion and location under different cameras. Experiment confirms that MFPCB allows pedestrian recognition to gain another round of performance boost. For instance, on the Mars dataset, we achieve 77.3% mAP and 88.6% rank-1 accuracy, surpassing the state of the art by a large margin.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116487963","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Combined Method Based on Source Text and Representation for Text Enhancement 基于源文本和表示的文本增强组合方法
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421519
Xuelian Li, Weihai Li, Yunxiao Zu
Text classification is a basic and important work in natural language processing (NLP). The existing text classification models are powerful. However, training such a model requires a large number of labeled training sets, but in the actual scene, insufficient data is often faced with. The lack of data is mainly divided into two categories: cold start and low resources. To solve this problem, text enhancement methods are usually used. In this paper, the source text enhancement and representation enhancement are combined to improve the enhancement effect. Five sets of experiments are designed to verify that our method is effective on different data sets and different classifiers. The simulation results show that the accuracy is improved and the generalization ability of the classifier is enhanced to some extent. We also find that the enhancement factor and the size of the training data set are not positively related to the enhancement effect. Therefore, the enhancement factor needs to be selected according to the characteristics of the data.
文本分类是自然语言处理(NLP)的一项基础和重要工作。现有的文本分类模型功能强大。然而,训练这样的模型需要大量的标记训练集,而在实际场景中,往往会面临数据不足的问题。数据缺失主要分为冷启动和资源不足两大类。为了解决这个问题,通常采用文本增强方法。本文采用源文本增强和表示增强相结合的方法来提高增强效果。设计了五组实验来验证我们的方法在不同数据集和不同分类器上的有效性。仿真结果表明,该方法在一定程度上提高了分类器的准确率和泛化能力。我们还发现增强因子和训练数据集的大小与增强效果并不是正相关的。因此,需要根据数据的特点来选择增强因子。
{"title":"Combined Method Based on Source Text and Representation for Text Enhancement","authors":"Xuelian Li, Weihai Li, Yunxiao Zu","doi":"10.1145/3421515.3421519","DOIUrl":"https://doi.org/10.1145/3421515.3421519","url":null,"abstract":"Text classification is a basic and important work in natural language processing (NLP). The existing text classification models are powerful. However, training such a model requires a large number of labeled training sets, but in the actual scene, insufficient data is often faced with. The lack of data is mainly divided into two categories: cold start and low resources. To solve this problem, text enhancement methods are usually used. In this paper, the source text enhancement and representation enhancement are combined to improve the enhancement effect. Five sets of experiments are designed to verify that our method is effective on different data sets and different classifiers. The simulation results show that the accuracy is improved and the generalization ability of the classifier is enhanced to some extent. We also find that the enhancement factor and the size of the training data set are not positively related to the enhancement effect. Therefore, the enhancement factor needs to be selected according to the characteristics of the data.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130957885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DeepFake Detection Algorithms: A Meta-Analysis DeepFake检测算法:一个元分析
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421532
Sergey Zotov, R. Dremliuga, A. Borshevnikov, Ksenia Krivosheeva
We analyzed the developed methods of computer vision in areas associated with recognition and detection of DeepFakes using various models and architectures of neural networks: mainly GAN and CNN. We also discussed the main types and models of these networks that are most effective in detecting and recognizing objects from different data sets, which were provided in the studied articles.
我们使用各种神经网络模型和架构(主要是GAN和CNN)分析了与DeepFakes识别和检测相关领域的计算机视觉开发方法。我们还讨论了这些网络的主要类型和模型,这些网络最有效地检测和识别来自不同数据集的目标,这些网络在研究文章中提供了。
{"title":"DeepFake Detection Algorithms: A Meta-Analysis","authors":"Sergey Zotov, R. Dremliuga, A. Borshevnikov, Ksenia Krivosheeva","doi":"10.1145/3421515.3421532","DOIUrl":"https://doi.org/10.1145/3421515.3421532","url":null,"abstract":"We analyzed the developed methods of computer vision in areas associated with recognition and detection of DeepFakes using various models and architectures of neural networks: mainly GAN and CNN. We also discussed the main types and models of these networks that are most effective in detecting and recognizing objects from different data sets, which were provided in the studied articles.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122986743","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Tyre Pattern Classification Based on Multi-scale GCN Model 基于多尺度GCN模型的轮胎花纹分类
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421520
Fuping Wang, Xiaoxia Ding, Y. Liu
Tyre pattern image classification plays an important role in traffic accidents and criminal scene investigation, and it contains rich texture structure information. Classic deep learning models, such as VGG, are often not targeted to represent the texture structure of tyre pattern images, and often cause over-fitting training due to large-scale parameters and insufficient training samples. To improve classification performance of tyre pattern image and solve the model overfitting problem, an efficient tyre pattern image classification model based on multi-scale Gabor convolutional neural network (MS-GCN) is proposed. First, a bank of large-scale directional Gabor filters are used to modulate the convolution kernel to extract more accurate texture features for large-size tyre pattern images, which greatly reduces the number of the training parameters and makes the model more streamlined. Secondly, due to the multi-scale texture similarity of the tyre pattern image, the multi-scale features from different convolutional layers are fused to produce the precise feature representation of the image, following by the optimal feature dimension selection. A large number of experiments were carried out on the real tyre pattern image data set. The results showed that the classification accuracy of the proposed algorithm is 95.9%, which is greatly improved compared with the handcrafted feature extraction algorithm and increased by 17.3% compared with deep learning-based model VGG16. In addition, the classification accuracy of the proposed algorithm on the GHIM-10K data set is 92%, which is also significantly improved compared to other methods. Overall, it shows the effectiveness and superiority of the proposed algorithm.
轮胎花纹图像分类包含丰富的纹理结构信息,在交通事故和犯罪现场侦查中具有重要作用。经典的深度学习模型,如VGG,往往没有针对性地表示轮胎图案图像的纹理结构,并且由于参数规模大,训练样本不足,往往导致训练过拟合。为了提高轮胎图案图像的分类性能,解决模型过拟合问题,提出了一种基于多尺度Gabor卷积神经网络(MS-GCN)的轮胎图案图像分类模型。首先,利用一组大规模定向Gabor滤波器对卷积核进行调制,对大尺寸轮胎图案图像提取更准确的纹理特征,大大减少了训练参数的数量,使模型更加精简;其次,利用轮胎图案图像的多尺度纹理相似性,融合不同卷积层的多尺度特征,得到图像的精确特征表示,并进行最优特征维数选择;在真实轮胎图案图像数据集上进行了大量的实验。结果表明,该算法的分类准确率为95.9%,与手工特征提取算法相比有很大提高,与基于深度学习的模型VGG16相比提高了17.3%。此外,本文算法在GHIM-10K数据集上的分类准确率为92%,与其他方法相比也有显著提高。总体而言,表明了该算法的有效性和优越性。
{"title":"Tyre Pattern Classification Based on Multi-scale GCN Model","authors":"Fuping Wang, Xiaoxia Ding, Y. Liu","doi":"10.1145/3421515.3421520","DOIUrl":"https://doi.org/10.1145/3421515.3421520","url":null,"abstract":"Tyre pattern image classification plays an important role in traffic accidents and criminal scene investigation, and it contains rich texture structure information. Classic deep learning models, such as VGG, are often not targeted to represent the texture structure of tyre pattern images, and often cause over-fitting training due to large-scale parameters and insufficient training samples. To improve classification performance of tyre pattern image and solve the model overfitting problem, an efficient tyre pattern image classification model based on multi-scale Gabor convolutional neural network (MS-GCN) is proposed. First, a bank of large-scale directional Gabor filters are used to modulate the convolution kernel to extract more accurate texture features for large-size tyre pattern images, which greatly reduces the number of the training parameters and makes the model more streamlined. Secondly, due to the multi-scale texture similarity of the tyre pattern image, the multi-scale features from different convolutional layers are fused to produce the precise feature representation of the image, following by the optimal feature dimension selection. A large number of experiments were carried out on the real tyre pattern image data set. The results showed that the classification accuracy of the proposed algorithm is 95.9%, which is greatly improved compared with the handcrafted feature extraction algorithm and increased by 17.3% compared with deep learning-based model VGG16. In addition, the classification accuracy of the proposed algorithm on the GHIM-10K data set is 92%, which is also significantly improved compared to other methods. Overall, it shows the effectiveness and superiority of the proposed algorithm.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131552722","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Color Recognition of Vehicle Based on Low Light Enhancement and Pixel-wise Contextual Attention 基于弱光增强和逐像素上下文注意的车辆颜色识别
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421527
Pengkang Zeng, JinTao Zhu, Guoheng Huang, Lianglun Cheng
At present, as a direction of intelligent transportation, the research results of car body color detection are still relatively lacking, and the current car body color detection is still easy to be affected by light, shielding, pollution and other factors. This paper proposes a color recognition of vehicle based on low light enhancement and Pixel-wise Contextual Attention, including low light intensity enhancement based on dual Fully Convolutional Networks (FCN), vehicle body detection based on Pixel-wise Contextual Attention Networks (PiCANet), and color classification of vehicle based on Convolutional Neural Network (CNN). The method of low light enhancement has better robustness and adaptability, and can better process the dark image. We use Pixel-wise Contextual Attention Networks, which better identify main area of vehicle with context information. Experiments show that our method is more accurate than the state-of-the-art method with 0.6% under insufficient light.
目前,作为智能交通的一个方向,车身颜色检测的研究成果还比较缺乏,目前的车身颜色检测还容易受到光线、遮挡、污染等因素的影响。本文提出了一种基于弱光增强和逐像素上下文注意的车辆颜色识别方法,包括基于双全卷积网络(FCN)的弱光增强、基于逐像素上下文注意网络(PiCANet)的车身检测和基于卷积神经网络(CNN)的车辆颜色分类。弱光增强方法具有较好的鲁棒性和适应性,能较好地处理暗图像。我们使用逐像素上下文注意网络,它可以更好地识别车辆的主要区域与上下文信息。实验表明,在光照不足的情况下,我们的方法比目前最先进的方法精度提高了0.6%。
{"title":"Color Recognition of Vehicle Based on Low Light Enhancement and Pixel-wise Contextual Attention","authors":"Pengkang Zeng, JinTao Zhu, Guoheng Huang, Lianglun Cheng","doi":"10.1145/3421515.3421527","DOIUrl":"https://doi.org/10.1145/3421515.3421527","url":null,"abstract":"At present, as a direction of intelligent transportation, the research results of car body color detection are still relatively lacking, and the current car body color detection is still easy to be affected by light, shielding, pollution and other factors. This paper proposes a color recognition of vehicle based on low light enhancement and Pixel-wise Contextual Attention, including low light intensity enhancement based on dual Fully Convolutional Networks (FCN), vehicle body detection based on Pixel-wise Contextual Attention Networks (PiCANet), and color classification of vehicle based on Convolutional Neural Network (CNN). The method of low light enhancement has better robustness and adaptability, and can better process the dark image. We use Pixel-wise Contextual Attention Networks, which better identify main area of vehicle with context information. Experiments show that our method is more accurate than the state-of-the-art method with 0.6% under insufficient light.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115359225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Image Correction Based on Field-Programmable Gate Array 基于现场可编程门阵列的图像校正
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421530
Xinrong Mao, Kaiming Liu
In machine vision, to correct the distortion of image is required. For improving the performance of the real-time distortion, this paper proposes an algorithm that can compress the inverse mapping table while conduct on-line reconstruction for the inverse mapping table by using interpolation method on FPGA platform in order to overcome the problems that FPGA, when be used to implement algorithm correcting image distortion, will become complexity in the on-line computation of the inverse mapping coordinate and perform insufficient in the capacity of on-chip ROM. The inverse mapping table is used to obtain inverse mapping coordinates that reduce both the amount of on-line computation of FPGA and the need of capacity of on chip ROM. The simulation on MATLAB show the results that when the compression parameters n is 4, 8, or 16, the distortion image can be corrected well and the information will not be lost. A FPGA-based double-sided visual image acquisition platform is built, and the algorithm is tested on the platform. Results show that the proposed algorithm can correct the nonlinear distortion of the image well.
在机器视觉中,需要对图像的畸变进行校正。为了提高图像实时失真的性能,本文提出了一种在FPGA平台上利用插值法对图像反映射表进行在线重构的同时对反映射表进行压缩的算法,以克服FPGA在实现图像畸变校正算法时存在的问题。将成为逆映射的在线计算复杂性协调和执行能力不足的片上罗逆映射表是用来获得逆映射坐标减少FPGA的在线计算量和产能的需要对芯片罗在MATLAB仿真结果表明当压缩参数n是4,8或16,失真图像可以纠正和信息不会丢失。搭建了基于fpga的双面视觉图像采集平台,并在平台上对算法进行了测试。结果表明,该算法能很好地校正图像的非线性畸变。
{"title":"Image Correction Based on Field-Programmable Gate Array","authors":"Xinrong Mao, Kaiming Liu","doi":"10.1145/3421515.3421530","DOIUrl":"https://doi.org/10.1145/3421515.3421530","url":null,"abstract":"In machine vision, to correct the distortion of image is required. For improving the performance of the real-time distortion, this paper proposes an algorithm that can compress the inverse mapping table while conduct on-line reconstruction for the inverse mapping table by using interpolation method on FPGA platform in order to overcome the problems that FPGA, when be used to implement algorithm correcting image distortion, will become complexity in the on-line computation of the inverse mapping coordinate and perform insufficient in the capacity of on-chip ROM. The inverse mapping table is used to obtain inverse mapping coordinates that reduce both the amount of on-line computation of FPGA and the need of capacity of on chip ROM. The simulation on MATLAB show the results that when the compression parameters n is 4, 8, or 16, the distortion image can be corrected well and the information will not be lost. A FPGA-based double-sided visual image acquisition platform is built, and the algorithm is tested on the platform. Results show that the proposed algorithm can correct the nonlinear distortion of the image well.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129756728","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Typicality of Lexical Bundles in Different Sections of Scientific Articles 科技文章不同章节词汇束的典型性
Pub Date : 2020-07-11 DOI: 10.1145/3421515.3421517
Haotong Wang, Y. Lepage, Chooi-Ling Goh
This paper proposes a method to quantify the typicality of lexical bundles in sections of academic articles, specifically in the field of Natural Language Processing papers. Typicality is defined as the product of individual KL-divergence scores and the probability of a bundle to appear in a type of section. An evaluation of our typicality measure against two other baselines shows slight improvements according to the Silhouette coefficient.
本文提出了一种量化学术文章章节中词汇束典型性的方法,特别是在自然语言处理领域的论文中。典型性被定义为个体kl -散度分数和一个束出现在一个类型截面中的概率的乘积。根据Silhouette系数,对我们的典型度量对另外两条基线的评估显示略有改善。
{"title":"Typicality of Lexical Bundles in Different Sections of Scientific Articles","authors":"Haotong Wang, Y. Lepage, Chooi-Ling Goh","doi":"10.1145/3421515.3421517","DOIUrl":"https://doi.org/10.1145/3421515.3421517","url":null,"abstract":"This paper proposes a method to quantify the typicality of lexical bundles in sections of academic articles, specifically in the field of Natural Language Processing papers. Typicality is defined as the product of individual KL-divergence scores and the probability of a bundle to appear in a type of section. An evaluation of our typicality measure against two other baselines shows slight improvements according to the Silhouette coefficient.","PeriodicalId":294293,"journal":{"name":"2020 2nd Symposium on Signal Processing Systems","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132144344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2020 2nd Symposium on Signal Processing Systems
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1