首页 > 最新文献

2021 13th International Conference on Machine Learning and Computing最新文献

英文 中文
Pedestrian Trajectory Prediction with MLP-social-GRU 基于MLP-social-GRU的行人轨迹预测
Pub Date : 2021-02-26 DOI: 10.1145/3457682.3457737
Yanbo Zhang, Liying Zheng
When crossing a crowded area, a person can predict dangers or collisions in advance around him/her, and then makes a suitable decision which direction he/she should take. The pedestrian trajectory prediction aims at simulating such ability of humans in a crowded environment. Most of the existing trajectory predictions are all based on the traditional hand-crafted methods that often ignore critical factors and can only be adapted to specific environments. Based on deep learning technology, this paper proposes a data-driven pedestrian trajectory predictor called MLP-social-GRU. First, the proposed predictor processes a pedestrian trajectory with a Multilayer Perceptron (MLP). Then, it adopts Gated Recurrent Units (GRU) to get hidden features of a pedestrian motion patterns, from which relationships between pedestrians can be simulated. Next, the social-pooling is used to receive and merge the hidden status information to get the mutual influence of adjacent pedestrians. Finally, a unified pedestrian trajectory prediction framework is designed based on abovementioned modules. We evaluate our predictor on two publicly available datasets, ETH and UCY, and the results show that it is superior to popular models such as LSTM, social-LSTM, and goal-social-array.
当穿过拥挤的区域时,人们可以提前预测周围的危险或碰撞,然后做出适当的决定,应该走哪个方向。行人轨迹预测就是为了模拟人类在拥挤环境中的这种能力。大多数现有的轨迹预测都是基于传统的手工制作的方法,往往忽略了关键因素,只能适应特定的环境。基于深度学习技术,提出了一种数据驱动的行人轨迹预测器MLP-social-GRU。首先,提出的预测器使用多层感知器(MLP)处理行人轨迹。然后,采用门控循环单元(GRU)获取行人运动模式的隐藏特征,以此模拟行人之间的关系;其次,利用social-pooling对隐藏的状态信息进行接收和合并,得到相邻行人的相互影响;最后,基于上述模块设计了统一的行人轨迹预测框架。我们在两个公开可用的数据集ETH和UCY上评估了我们的预测器,结果表明它优于流行的模型,如LSTM, social-LSTM和goal-social-array。
{"title":"Pedestrian Trajectory Prediction with MLP-social-GRU","authors":"Yanbo Zhang, Liying Zheng","doi":"10.1145/3457682.3457737","DOIUrl":"https://doi.org/10.1145/3457682.3457737","url":null,"abstract":"When crossing a crowded area, a person can predict dangers or collisions in advance around him/her, and then makes a suitable decision which direction he/she should take. The pedestrian trajectory prediction aims at simulating such ability of humans in a crowded environment. Most of the existing trajectory predictions are all based on the traditional hand-crafted methods that often ignore critical factors and can only be adapted to specific environments. Based on deep learning technology, this paper proposes a data-driven pedestrian trajectory predictor called MLP-social-GRU. First, the proposed predictor processes a pedestrian trajectory with a Multilayer Perceptron (MLP). Then, it adopts Gated Recurrent Units (GRU) to get hidden features of a pedestrian motion patterns, from which relationships between pedestrians can be simulated. Next, the social-pooling is used to receive and merge the hidden status information to get the mutual influence of adjacent pedestrians. Finally, a unified pedestrian trajectory prediction framework is designed based on abovementioned modules. We evaluate our predictor on two publicly available datasets, ETH and UCY, and the results show that it is superior to popular models such as LSTM, social-LSTM, and goal-social-array.","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132772021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Diverse Conversation Generation System with Sentence Function Classification 具有句子功能分类的多元会话生成系统
Pub Date : 2021-02-26 DOI: 10.1145/3457682.3457761
Zuning Fan, Liangwei Chen
This paper mainly studies the implementation of diverse conversation generation system based on end-to-end neural networks and sentence classification on sentence function. In existing work, the output of the conversational system is mostly in the form of one type of sentence, for instance, declarative sentence. The type of sentence function is not well distributed. There is still insufficient diversity in the output of conversational system, which could be unattractive to users. A good conversational system could well interact with users by generating diverse output, including asking and responding, driving conversations to go further. Generating different type of output sentence is necessary to conversational systems, which is also challenging. Therefore, this paper introduces the idea of diverse conversation into the generative system, and designs the Diverse Conversation Generation (DCG) model. The model adopts a sentence function classifier trained independently to supervise the model output with modified loss function and back-propagation. The DCG model increases the diversity of output sentence, which could guide user to chat more with the system, extend the quality of conversation, and improve the user experience. The model is experimented on two different sequence-to sequence models, evaluated with perplexity and classify entropy, achieves better performance compared with two base models.
本文主要研究了基于端到端神经网络和基于句子功能的句子分类的多样化会话生成系统的实现。在现有的工作中,会话系统的输出大多是一种句子的形式,例如陈述句。句子功能的类型分布不均匀。会话系统的输出仍然缺乏多样性,这可能对用户没有吸引力。一个好的会话系统可以很好地与用户进行交互,通过生成不同的输出,包括提问和回应,推动对话进一步发展。会话系统需要生成不同类型的输出句子,这也是一个挑战。为此,本文将多元对话的思想引入生成系统,设计了多元对话生成(DCG)模型。该模型采用独立训练的句子函数分类器,通过修正损失函数和反向传播对模型输出进行监督。DCG模型增加了输出句子的多样性,可以引导用户更多地与系统聊天,扩展对话质量,改善用户体验。该模型在两种不同的序列对序列模型上进行了实验,用困惑度和分类熵进行了评价,与两种基本模型相比,获得了更好的性能。
{"title":"Diverse Conversation Generation System with Sentence Function Classification","authors":"Zuning Fan, Liangwei Chen","doi":"10.1145/3457682.3457761","DOIUrl":"https://doi.org/10.1145/3457682.3457761","url":null,"abstract":"This paper mainly studies the implementation of diverse conversation generation system based on end-to-end neural networks and sentence classification on sentence function. In existing work, the output of the conversational system is mostly in the form of one type of sentence, for instance, declarative sentence. The type of sentence function is not well distributed. There is still insufficient diversity in the output of conversational system, which could be unattractive to users. A good conversational system could well interact with users by generating diverse output, including asking and responding, driving conversations to go further. Generating different type of output sentence is necessary to conversational systems, which is also challenging. Therefore, this paper introduces the idea of diverse conversation into the generative system, and designs the Diverse Conversation Generation (DCG) model. The model adopts a sentence function classifier trained independently to supervise the model output with modified loss function and back-propagation. The DCG model increases the diversity of output sentence, which could guide user to chat more with the system, extend the quality of conversation, and improve the user experience. The model is experimented on two different sequence-to sequence models, evaluated with perplexity and classify entropy, achieves better performance compared with two base models.","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115716901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
AEE-Net: An Efficient End-to-End Dehazing Network in UAV Imaging System AEE-Net:无人机成像系统中高效的端到端除雾网络
Pub Date : 2021-02-26 DOI: 10.1145/3457682.3457739
Tianxiao Cai, Sheng Zhang, Bo Tan
Because it can provide real-time images for the first time, UAV plays a massive role in disaster relief, environmental observation, and information collection. However, the quality of images collected by UAV is always affected by fog. Therefore, the research on how to remove the fog in the image becomes more and more critical. In recent years, the role of convolutional neural networks (CNN), which can automatically extract features and efficiently process high-dimensional data, has received more and more attention in many disciplines. To improve the imaging quality of UAV in a foggy environment, this paper proposes an image dehazing model built with a convolutional neural network (CNN), called an effective end-to-end dehazing Network (AEE-Net). Our proposed method has a faster running speed than traditional models due to the simple structure of the model and the design based on the modified atmospheric scattering model. Our method combines the characteristics of dehazing processes and the advantages of deep learning. Experimental results on the training set and raw images show that the proposed method has better performance than traditional methods. This method can improve the quality of UAV-captured images under foggy conditions and can meet the input requirements of UAV vision tasks.
因为它可以第一次提供实时图像,无人机在救灾、环境观测和信息收集方面发挥了巨大的作用。然而,无人机采集的图像质量总是受到雾的影响。因此,如何去除图像中的雾的研究变得越来越重要。近年来,卷积神经网络(convolutional neural network, CNN)以其自动提取特征和高效处理高维数据的能力,在许多学科中受到越来越多的关注。为了提高多雾环境下无人机的成像质量,本文提出了一种基于卷积神经网络(CNN)的图像去雾模型,称为有效端到端去雾网络(AEE-Net)。由于模型结构简单,且基于改进的大气散射模型设计,因此该方法比传统模型运行速度更快。我们的方法结合了除雾过程的特点和深度学习的优点。在训练集和原始图像上的实验结果表明,该方法比传统方法具有更好的性能。该方法可以提高无人机在雾天条件下捕获图像的质量,满足无人机视觉任务的输入要求。
{"title":"AEE-Net: An Efficient End-to-End Dehazing Network in UAV Imaging System","authors":"Tianxiao Cai, Sheng Zhang, Bo Tan","doi":"10.1145/3457682.3457739","DOIUrl":"https://doi.org/10.1145/3457682.3457739","url":null,"abstract":"Because it can provide real-time images for the first time, UAV plays a massive role in disaster relief, environmental observation, and information collection. However, the quality of images collected by UAV is always affected by fog. Therefore, the research on how to remove the fog in the image becomes more and more critical. In recent years, the role of convolutional neural networks (CNN), which can automatically extract features and efficiently process high-dimensional data, has received more and more attention in many disciplines. To improve the imaging quality of UAV in a foggy environment, this paper proposes an image dehazing model built with a convolutional neural network (CNN), called an effective end-to-end dehazing Network (AEE-Net). Our proposed method has a faster running speed than traditional models due to the simple structure of the model and the design based on the modified atmospheric scattering model. Our method combines the characteristics of dehazing processes and the advantages of deep learning. Experimental results on the training set and raw images show that the proposed method has better performance than traditional methods. This method can improve the quality of UAV-captured images under foggy conditions and can meet the input requirements of UAV vision tasks.","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"94 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116317446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Using Deep Learning to Construct Auto Web Penetration Test 利用深度学习构建汽车Web渗透测试
Pub Date : 2021-02-26 DOI: 10.1145/3457682.3457691
Jian Jiao, Haini Zhao, Hongsheng Cao
Penetration test is an important means to test the security of the web system. It has been mainly carried out by tester manually. The main reason is that it is difficult to generate test path and code automatically because of the complex network environment. The traditional method for attack path can't give the code for the whole penetration process. The traditional penetration path is based on the correlation between vulnerabilities and lacks practical experience support. In this paper, we propose a method based on CNN, which can automatically produce the code of penetration test by training the data which originate from the real attack events. We further implement the system to verify it. In a real environment experiment, we have validated the system, and analyzed the feasibility and performance of the CNN technology for penetration tests.
渗透测试是测试web系统安全性的重要手段。主要由测试人员手工完成。主要原因是由于复杂的网络环境,难以自动生成测试路径和代码。传统的攻击路径方法无法给出整个渗透过程的代码。传统的渗透路径是基于漏洞之间的相关性,缺乏实际经验支持。本文提出了一种基于CNN的方法,通过训练来源于真实攻击事件的数据,自动生成渗透测试代码。我们进一步实施该系统来验证它。在真实环境实验中,我们对系统进行了验证,并分析了CNN技术用于突防测试的可行性和性能。
{"title":"Using Deep Learning to Construct Auto Web Penetration Test","authors":"Jian Jiao, Haini Zhao, Hongsheng Cao","doi":"10.1145/3457682.3457691","DOIUrl":"https://doi.org/10.1145/3457682.3457691","url":null,"abstract":"Penetration test is an important means to test the security of the web system. It has been mainly carried out by tester manually. The main reason is that it is difficult to generate test path and code automatically because of the complex network environment. The traditional method for attack path can't give the code for the whole penetration process. The traditional penetration path is based on the correlation between vulnerabilities and lacks practical experience support. In this paper, we propose a method based on CNN, which can automatically produce the code of penetration test by training the data which originate from the real attack events. We further implement the system to verify it. In a real environment experiment, we have validated the system, and analyzed the feasibility and performance of the CNN technology for penetration tests.","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"271 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115997677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Using Video Restoration to Improve Face Forgery Detection Based on Low-quality Video 基于低质量视频的视频复原改进人脸伪造检测
Pub Date : 2021-02-26 DOI: 10.1145/3457682.3457753
Jianyang Qi, Peng Liang, Gang Hao, Yuting Wu
The face forgery detection model based on XceptionNet has made great achievements in the detection field. However, it is still a challenge to detect fake faces in low-quality video images, because the low-quality video image has an insufficient resolution, which leads to the loss of details of the video image, thus resulting in image blurring. To solve this problem, this paper proposes a low-quality video face forgery detection method based on video recovery. This method mainly uses the Pyramid, Cascading, and Deformable convolution(PCD) module and the spatiotemporal attention (TSA) fusion module to restore low-quality video face images, and then obtains the restored feature map. And then the restored feature map is fed into the Xception classification network for face forgery detection. Moreover, The pre-training model parameters based on ImageNet makes the training model converge on 2GPU days. The results show that this method has a good experimental effect on the test data set.
基于XceptionNet的人脸伪造检测模型在检测领域取得了很大的成就。然而,在低质量视频图像中检测假脸仍然是一个挑战,因为低质量视频图像的分辨率不足,导致视频图像的细节丢失,从而导致图像模糊。为了解决这一问题,本文提出了一种基于视频恢复的低质量视频人脸伪造检测方法。该方法主要利用金字塔级联变形卷积(Pyramid, Cascading, and Deformable convolution, PCD)模块和时空注意力(spatial - temporal attention, TSA)融合模块对低质量视频人脸图像进行恢复,然后得到恢复后的特征图。然后将恢复的特征图输入异常分类网络进行人脸伪造检测。此外,基于ImageNet的预训练模型参数使训练模型在2GPU天内收敛。结果表明,该方法在测试数据集上具有良好的实验效果。
{"title":"Using Video Restoration to Improve Face Forgery Detection Based on Low-quality Video","authors":"Jianyang Qi, Peng Liang, Gang Hao, Yuting Wu","doi":"10.1145/3457682.3457753","DOIUrl":"https://doi.org/10.1145/3457682.3457753","url":null,"abstract":"The face forgery detection model based on XceptionNet has made great achievements in the detection field. However, it is still a challenge to detect fake faces in low-quality video images, because the low-quality video image has an insufficient resolution, which leads to the loss of details of the video image, thus resulting in image blurring. To solve this problem, this paper proposes a low-quality video face forgery detection method based on video recovery. This method mainly uses the Pyramid, Cascading, and Deformable convolution(PCD) module and the spatiotemporal attention (TSA) fusion module to restore low-quality video face images, and then obtains the restored feature map. And then the restored feature map is fed into the Xception classification network for face forgery detection. Moreover, The pre-training model parameters based on ImageNet makes the training model converge on 2GPU days. The results show that this method has a good experimental effect on the test data set.","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"8 16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129137716","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Toward an Effective Analysis of COVID-19 Moroccan Business Survey Data using Machine Learning Techniques 利用机器学习技术有效分析COVID-19摩洛哥商业调查数据
Pub Date : 2021-02-26 DOI: 10.1145/3457682.3457690
Imane Lasri, Anouar Riadsolh, Mourad Elbelkacemi
COVID-19 pandemic has gravely affected our societies and economies with severe consequences. To contain the spread of the disease, most governments around the world authorized unprecedented measures, including Morocco, which has closed the borders and adopted full lockdown between March and June 2020. However, these measures have resulted in economic loss and have led to dramatic changes in how businesses act and consumers behave. The main focus of this study was to examine the impact of the full lockdown on Moroccan enterprises based on the COVID-19 Moroccan business survey carried out by the High Commission for Planning (HCP). A three-stage analysis method was employed. First, multiple correspondence analysis (MCA) was used to reduce the dimensionality of the categorical variables, and k-means clustering algorithm was used to cluster the data, then decision tree algorithm was performed in order to interpret each cluster and the maximum accuracy achieved is 84.45%. Compared with the decision tree algorithm, an artificial neural network (ANN) with stratified 10-fold cross-validation was applied to the dataset and has reached an accuracy of 83.4%. The simulation results confirm the effectiveness of the proposed techniques for analyzing survey data.
COVID-19大流行严重影响了我们的社会和经济,造成了严重后果。为了控制疾病的传播,世界上大多数国家的政府都采取了前所未有的措施,包括摩洛哥,该国在2020年3月至6月期间关闭了边境,并采取了全面封锁措施。然而,这些措施造成了经济损失,并导致了企业行为和消费者行为的巨大变化。本研究的主要重点是根据规划高级委员会(HCP)开展的新冠肺炎摩洛哥商业调查,研究全面封锁对摩洛哥企业的影响。采用三阶段分析法。首先利用多重对应分析(multiple correspondence analysis, MCA)对分类变量进行降维,然后利用k-means聚类算法对数据进行聚类,然后利用决策树算法对每一聚类进行解释,最高准确率达到84.45%。与决策树算法相比,采用分层10倍交叉验证的人工神经网络(ANN)对数据集进行分析,准确率达到83.4%。仿真结果验证了所提方法对测量数据分析的有效性。
{"title":"Toward an Effective Analysis of COVID-19 Moroccan Business Survey Data using Machine Learning Techniques","authors":"Imane Lasri, Anouar Riadsolh, Mourad Elbelkacemi","doi":"10.1145/3457682.3457690","DOIUrl":"https://doi.org/10.1145/3457682.3457690","url":null,"abstract":"COVID-19 pandemic has gravely affected our societies and economies with severe consequences. To contain the spread of the disease, most governments around the world authorized unprecedented measures, including Morocco, which has closed the borders and adopted full lockdown between March and June 2020. However, these measures have resulted in economic loss and have led to dramatic changes in how businesses act and consumers behave. The main focus of this study was to examine the impact of the full lockdown on Moroccan enterprises based on the COVID-19 Moroccan business survey carried out by the High Commission for Planning (HCP). A three-stage analysis method was employed. First, multiple correspondence analysis (MCA) was used to reduce the dimensionality of the categorical variables, and k-means clustering algorithm was used to cluster the data, then decision tree algorithm was performed in order to interpret each cluster and the maximum accuracy achieved is 84.45%. Compared with the decision tree algorithm, an artificial neural network (ANN) with stratified 10-fold cross-validation was applied to the dataset and has reached an accuracy of 83.4%. The simulation results confirm the effectiveness of the proposed techniques for analyzing survey data.","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130571348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Inter-domain Link Inference with Confidence Using Naïve Bayes Classifier 使用Naïve贝叶斯分类器的域间链接推理
Pub Date : 2021-02-26 DOI: 10.1145/3457682.3457715
Yi Zhao, Yan Liu, Xiaoyu Guo, ZhongHang Sui
Inter-domain link inference is not only important for network security and fault diagnosis, but also helps to conduct research on inter-domain congestion detection and network resilience assessment. Current researches on this issue lack confidence analysis of the inferred results. In this paper, the IP link types (i.e., intra-domain link and inter-domain link) are considered as the latent variable in probability model, while the parameters are probabilities of different link types with particular features. The expectation maximization algorithm is applied to estimate parameters of the model. In each iteration of EM algorithm, Naïve Bayes is used for classification. The final result is determined according to the probability, and the probability is the confidence of the result. The experimental results show that our method can achieve better precision and recall on the validation set than two existing general methods.
域间链路推理不仅对网络安全和故障诊断具有重要意义,而且有助于进行域间拥塞检测和网络弹性评估的研究。目前对这一问题的研究缺乏对推断结果的置信度分析。本文将IP链路类型(即域内链路和域间链路)作为概率模型的潜在变量,参数为具有特定特征的不同链路类型的概率。采用期望最大化算法对模型参数进行估计。在EM算法的每次迭代中,都使用Naïve贝叶斯进行分类。最终的结果是根据概率决定的,而概率就是对结果的置信度。实验结果表明,该方法在验证集上的查全率和查全率优于现有的两种一般方法。
{"title":"Inter-domain Link Inference with Confidence Using Naïve Bayes Classifier","authors":"Yi Zhao, Yan Liu, Xiaoyu Guo, ZhongHang Sui","doi":"10.1145/3457682.3457715","DOIUrl":"https://doi.org/10.1145/3457682.3457715","url":null,"abstract":"Inter-domain link inference is not only important for network security and fault diagnosis, but also helps to conduct research on inter-domain congestion detection and network resilience assessment. Current researches on this issue lack confidence analysis of the inferred results. In this paper, the IP link types (i.e., intra-domain link and inter-domain link) are considered as the latent variable in probability model, while the parameters are probabilities of different link types with particular features. The expectation maximization algorithm is applied to estimate parameters of the model. In each iteration of EM algorithm, Naïve Bayes is used for classification. The final result is determined according to the probability, and the probability is the confidence of the result. The experimental results show that our method can achieve better precision and recall on the validation set than two existing general methods.","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117013761","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Single-Pass On-Line Event Detection in Twitter Streams Twitter流中的单次在线事件检测
Pub Date : 2021-02-26 DOI: 10.1145/3457682.3457762
Xingfa Qiu, Qiaosha Zou, C. Richard Shi
Intensive information is emerged in social media every second. Many breaking news often appear first in social media, much earlier than they appear in traditional news media. Through the technology of event detection on social media data streams, scatter information can be gathered together to inform us the popular events discussing online. An event is often modeled as a cluster of documents which discuss the same subject. Traditional event detection methods perform poorly on social media because of their huge amount of data and irregular expressions. In this paper, we propose a simple yet efficient event detection method towards social media. An event is represented by a sequence of keywords extracted from social media. We use a single-pass incremental clustering method with a trained encoder mapping documents and events into the same semantic space, which is helpful for the similarity calculation between them. We consider the similarity calculation between a tweet and an event as a matching process and construct a relevance matching dataset with tweet-event pairs. We finetune BERT (Bidirectional Encoder Representations from Transformers) model in the matching dataset to get an appropriate semantic encoder. Keywords are dynamically changed to represent an event for capturing the development of the event. Our proposed method achieves 0.86 on NMI (Normed Mutual Information), 0.69 on ARI (Adjusted Rand Index) and 0.70 on F1-score on a public twitter dataset, which shows the superiority of our method compared with baseline methods.
社交媒体上每秒钟都有大量信息涌现。许多突发新闻往往首先出现在社交媒体上,比传统新闻媒体出现的时间要早得多。通过对社交媒体数据流的事件检测技术,将分散的信息收集在一起,告诉我们网络上讨论的热门事件。事件通常被建模为讨论同一主题的一组文档。传统的事件检测方法在社交媒体上表现不佳,因为社交媒体的数据量大,表达式不规则。在本文中,我们提出了一种简单而高效的针对社交媒体的事件检测方法。事件由从社交媒体中提取的一系列关键字表示。我们使用一种单遍增量聚类方法,通过训练好的编码器将文档和事件映射到相同的语义空间,这有助于它们之间的相似度计算。我们将推文和事件之间的相似度计算视为一个匹配过程,并使用推文-事件对构建一个相关匹配数据集。我们在匹配数据集中对BERT (Bidirectional Encoder Representations from Transformers)模型进行微调,得到一个合适的语义编码器。动态更改关键字以表示事件,以便捕获事件的发展。我们提出的方法在NMI (normmed Mutual Information)上达到0.86,在ARI (Adjusted Rand Index)上达到0.69,在F1-score上达到0.70,这表明我们的方法与基线方法相比具有优势。
{"title":"Single-Pass On-Line Event Detection in Twitter Streams","authors":"Xingfa Qiu, Qiaosha Zou, C. Richard Shi","doi":"10.1145/3457682.3457762","DOIUrl":"https://doi.org/10.1145/3457682.3457762","url":null,"abstract":"Intensive information is emerged in social media every second. Many breaking news often appear first in social media, much earlier than they appear in traditional news media. Through the technology of event detection on social media data streams, scatter information can be gathered together to inform us the popular events discussing online. An event is often modeled as a cluster of documents which discuss the same subject. Traditional event detection methods perform poorly on social media because of their huge amount of data and irregular expressions. In this paper, we propose a simple yet efficient event detection method towards social media. An event is represented by a sequence of keywords extracted from social media. We use a single-pass incremental clustering method with a trained encoder mapping documents and events into the same semantic space, which is helpful for the similarity calculation between them. We consider the similarity calculation between a tweet and an event as a matching process and construct a relevance matching dataset with tweet-event pairs. We finetune BERT (Bidirectional Encoder Representations from Transformers) model in the matching dataset to get an appropriate semantic encoder. Keywords are dynamically changed to represent an event for capturing the development of the event. Our proposed method achieves 0.86 on NMI (Normed Mutual Information), 0.69 on ARI (Adjusted Rand Index) and 0.70 on F1-score on a public twitter dataset, which shows the superiority of our method compared with baseline methods.","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"88 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127207776","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
ICodeNet - A Hierarchical Neural Network Approach For Source Code Author Identification 源代码作者识别的层次神经网络方法
Pub Date : 2021-01-30 DOI: 10.1145/3457682.3457709
Pranali Bora, Tulika Awalgaonkar, Himanshu Palve, Raviraj Joshi, Purvi Goel
With the open-source revolution, source codes are now more easily accessible than ever. This has, however, made it easier for malicious users and institutions to copy the code without giving regards to the license, or credit to the original author. Therefore, source code author identification is a critical task with paramount importance. In this paper, we propose ICodeNet - a hierarchical neural network that can be used for source code file-level tasks. The ICodeNet processes source code in image format and is employed for the task of per file author identification. The ICodeNet consists of an ImageNet trained VGG encoder followed by a shallow neural network. The shallow network is based either on CNN or LSTM. Different variations of models are evaluated on a source code author classification dataset. We have also compared our image-based hierarchical neural network model with simple image-based CNN architecture and text-based CNN and LSTM models to highlight its novelty and efficiency.
随着开放源代码的革命,源代码现在比以往任何时候都更容易获得。然而,这使得恶意用户和机构更容易复制代码,而无需考虑许可证或原始作者的信用。因此,源代码作者识别是一项至关重要的关键任务。在本文中,我们提出了ICodeNet -一个可用于源代码文件级任务的分层神经网络。ICodeNet以图像格式处理源代码,并用于每个文件作者识别的任务。ICodeNet由ImageNet训练的VGG编码器和浅神经网络组成。浅层网络基于CNN或LSTM。在源代码作者分类数据集上评估模型的不同变体。我们还将基于图像的分层神经网络模型与简单的基于图像的CNN架构和基于文本的CNN和LSTM模型进行了比较,以突出其新颖性和效率。
{"title":"ICodeNet - A Hierarchical Neural Network Approach For Source Code Author Identification","authors":"Pranali Bora, Tulika Awalgaonkar, Himanshu Palve, Raviraj Joshi, Purvi Goel","doi":"10.1145/3457682.3457709","DOIUrl":"https://doi.org/10.1145/3457682.3457709","url":null,"abstract":"With the open-source revolution, source codes are now more easily accessible than ever. This has, however, made it easier for malicious users and institutions to copy the code without giving regards to the license, or credit to the original author. Therefore, source code author identification is a critical task with paramount importance. In this paper, we propose ICodeNet - a hierarchical neural network that can be used for source code file-level tasks. The ICodeNet processes source code in image format and is employed for the task of per file author identification. The ICodeNet consists of an ImageNet trained VGG encoder followed by a shallow neural network. The shallow network is based either on CNN or LSTM. Different variations of models are evaluated on a source code author classification dataset. We have also compared our image-based hierarchical neural network model with simple image-based CNN architecture and text-based CNN and LSTM models to highlight its novelty and efficiency.","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-01-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126228653","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Content-Based Textual File Type Detection at Scale 大规模基于内容的文本文件类型检测
Pub Date : 2021-01-21 DOI: 10.1145/3457682.3457756
Francesca Del Bonifro, M. Gabbrielli, Stefano Zacchiroli
Programming language detection is a common need in the analysis of large source code bases. It is supported by a number of existing tools that rely on several features, and most notably file extensions, to determine file types. We consider the problem of accurately detecting the type of files commonly found in software code bases, based solely on textual file content. Doing so is helpful to classify source code that lack file extensions (e.g., code snippets posted on the Web or executable scripts), to avoid misclassifying source code that has been recorded with wrong or uncommon file extensions, and also shed some light on the intrinsic recognizability of source code files. We propose a simple model that (a) use a language-agnostic word tokenizer for textual files, (b) group tokens in 1-/2-grams, (c) build feature vectors based on N-gram frequencies, and (d) use a simple fully connected neural network as classifier. As training set we use textual files extracted from GitHub repositories with at least 1000 stars, using existing file extensions as ground truth. Despite its simplicity the proposed model reaches ≈ 85% in our experiments for a relatively high number of recognized classes (more than 130 file types).
在分析大型源代码库时,编程语言检测是一种常见的需求。许多现有的工具都支持它,这些工具依赖于几个特性,尤其是文件扩展名来确定文件类型。我们考虑的问题是准确地检测软件代码库中常见的文件类型,仅基于文本文件内容。这样做有助于对缺乏文件扩展名的源代码进行分类(例如,发布在Web上的代码片段或可执行脚本),以避免对使用错误或不常见的文件扩展名记录的源代码进行错误分类,并且还揭示了源代码文件的内在可识别性。我们提出了一个简单的模型:(a)对文本文件使用语言无关的词标记器,(b)将标记按1-/2-g分组,(c)基于N-gram频率构建特征向量,(d)使用简单的全连接神经网络作为分类器。作为训练集,我们使用从GitHub存储库中提取的文本文件,至少有1000个星星,使用现有的文件扩展名作为基础事实。尽管它很简单,但在我们的实验中,对于相对较多的可识别类(超过130个文件类型),所提出的模型达到了≈85%。
{"title":"Content-Based Textual File Type Detection at Scale","authors":"Francesca Del Bonifro, M. Gabbrielli, Stefano Zacchiroli","doi":"10.1145/3457682.3457756","DOIUrl":"https://doi.org/10.1145/3457682.3457756","url":null,"abstract":"Programming language detection is a common need in the analysis of large source code bases. It is supported by a number of existing tools that rely on several features, and most notably file extensions, to determine file types. We consider the problem of accurately detecting the type of files commonly found in software code bases, based solely on textual file content. Doing so is helpful to classify source code that lack file extensions (e.g., code snippets posted on the Web or executable scripts), to avoid misclassifying source code that has been recorded with wrong or uncommon file extensions, and also shed some light on the intrinsic recognizability of source code files. We propose a simple model that (a) use a language-agnostic word tokenizer for textual files, (b) group tokens in 1-/2-grams, (c) build feature vectors based on N-gram frequencies, and (d) use a simple fully connected neural network as classifier. As training set we use textual files extracted from GitHub repositories with at least 1000 stars, using existing file extensions as ground truth. Despite its simplicity the proposed model reaches ≈ 85% in our experiments for a relatively high number of recognized classes (more than 130 file types).","PeriodicalId":142045,"journal":{"name":"2021 13th International Conference on Machine Learning and Computing","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-01-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116766747","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2021 13th International Conference on Machine Learning and Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1