首页 > 最新文献

2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)最新文献

英文 中文
[Title page i] [标题页i]
{"title":"[Title page i]","authors":"","doi":"10.1109/ictai.2019.00001","DOIUrl":"https://doi.org/10.1109/ictai.2019.00001","url":null,"abstract":"","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"276 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131861228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An Advanced Harmony Search Algorithm Based on Harmony Anchoring and Reverse Learning 一种基于和谐锚定和逆向学习的高级和谐搜索算法
Lin Liu, D. Shi, Dansong Cheng, Maysam Orouskhani
In this paper, we propose a new and effective multi-objective optimization algorithm based on a modified harmony search. The proposed method employs reverse learning in the harmony vector updating equation in order to enhance the global searching ability. Moreover, it adopts a harmony anchoring scheme so that unnecessary exploration is avoided. Experimental studies carried on eight benchmark problems show quite satisfactory results and indicate the higher performance of the proposed algorithm in comparison with traditional multi-objective optimization algorithms. Finally, it has been applied to solve the image segmentation problem.
本文提出了一种新的、有效的基于改进和声搜索的多目标优化算法。该方法在和声向量更新方程中采用逆向学习,增强了全局搜索能力。并且采用和谐锚定方案,避免了不必要的探索。对8个基准问题进行了实验研究,结果令人满意,与传统的多目标优化算法相比,该算法具有更高的性能。最后,将其应用于解决图像分割问题。
{"title":"An Advanced Harmony Search Algorithm Based on Harmony Anchoring and Reverse Learning","authors":"Lin Liu, D. Shi, Dansong Cheng, Maysam Orouskhani","doi":"10.1109/ICTAI.2019.00255","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00255","url":null,"abstract":"In this paper, we propose a new and effective multi-objective optimization algorithm based on a modified harmony search. The proposed method employs reverse learning in the harmony vector updating equation in order to enhance the global searching ability. Moreover, it adopts a harmony anchoring scheme so that unnecessary exploration is avoided. Experimental studies carried on eight benchmark problems show quite satisfactory results and indicate the higher performance of the proposed algorithm in comparison with traditional multi-objective optimization algorithms. Finally, it has been applied to solve the image segmentation problem.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132375785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
BERT for Stock Market Sentiment Analysis 股票市场情绪分析
Matheus Gomes Sousa, K. Sakiyama, Lucas de Souza Rodrigues, Pedro Henrique de Moraes, Eraldo Rezende Fernandes, E. Matsubara
When breaking news occurs, stock quotes can change abruptly in a matter of seconds. The human analysis of breaking news can take several minutes, and investors in the financial markets need to make quick decisions. Such challenging scenarios require faster ways to support investors. In this work, we propose the use of bidirectional encoder representations from transformers BERT to perform sentiment analysis of news articles and provide relevant information for decision making in the stock market. This model is pre-trained on a large amount of general-domain documents by means of a self-learning task. To fine-tune this powerful model on sentiment analysis for the stock market, we manually labeled stock news articles as positive, neutral or negative. This dataset is freely available and amounts to 582 documents from several financial news sources. We fine-tune a BERT model on this dataset and achieve 72.5% of F-score. Then, we perform some experiments highlighting how the output of the obtained model can provide valuable information to predict the subsequent movements of the Dow Jones Industrial (DJI) Index.
当突发新闻发生时,股票报价可能在几秒钟内突然发生变化。人类对突发新闻的分析可能需要几分钟,而金融市场的投资者需要迅速做出决定。这种具有挑战性的情况需要更快的方式来支持投资者。在这项工作中,我们建议使用来自变压器BERT的双向编码器表示来执行新闻文章的情感分析,并为股票市场的决策提供相关信息。该模型通过自学习任务在大量通用领域文档上进行预训练。为了对这个强大的股市情绪分析模型进行微调,我们手动将股票新闻文章标记为积极、中性或消极。这个数据集是免费提供的,总共有582份来自几个金融新闻来源的文件。我们对该数据集的BERT模型进行了微调,达到了72.5%的f值。然后,我们进行了一些实验,突出了所获得的模型的输出如何为预测道琼斯工业指数(DJI)的后续走势提供有价值的信息。
{"title":"BERT for Stock Market Sentiment Analysis","authors":"Matheus Gomes Sousa, K. Sakiyama, Lucas de Souza Rodrigues, Pedro Henrique de Moraes, Eraldo Rezende Fernandes, E. Matsubara","doi":"10.1109/ICTAI.2019.00231","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00231","url":null,"abstract":"When breaking news occurs, stock quotes can change abruptly in a matter of seconds. The human analysis of breaking news can take several minutes, and investors in the financial markets need to make quick decisions. Such challenging scenarios require faster ways to support investors. In this work, we propose the use of bidirectional encoder representations from transformers BERT to perform sentiment analysis of news articles and provide relevant information for decision making in the stock market. This model is pre-trained on a large amount of general-domain documents by means of a self-learning task. To fine-tune this powerful model on sentiment analysis for the stock market, we manually labeled stock news articles as positive, neutral or negative. This dataset is freely available and amounts to 582 documents from several financial news sources. We fine-tune a BERT model on this dataset and achieve 72.5% of F-score. Then, we perform some experiments highlighting how the output of the obtained model can provide valuable information to predict the subsequent movements of the Dow Jones Industrial (DJI) Index.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"49 19","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132390038","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
TFPN: Twin Feature Pyramid Networks for Object Detection TFPN:用于目标检测的双特征金字塔网络
Yi Liang, Changjian Wang, Fangzhao Li, Yuxing Peng, Q. Lv, Yuan Yuan, Zhen Huang
FPN (Feature Pyramid Networks) is one of the most popular object detection networks, which can improve small object detection by enhancing shallow features. However, limited attention has been paid to the improvement of large object detection via deeper feature enhancement. One existing approach merges the feature maps of different layers into a new feature map for object detection, but can lead to increased noise and loss of information. The other approach adds a bottom-up structure after the feature pyramid of FPN, which superimposes the information from shallow layers into the deep feature map but weakens the strength of FPN in detecting small objects. To address these challenges, this paper proposes TFPN (Twin Feature Pyramid Networks), which consists of (1) FPN+, a bottom-up structure that improves large object detection; (2) TPS, a Twin Pyramid Structure that improves medium object detection; and (3) innovative integration of these two with FPN, which can significantly improve the detection accuracy of large and medium objects while maintaining the advantage of FPN in small object detection. Extensive experiments using the MSCOCO object detection datasets and the BDD100K automatic driving dataset demonstrate that TFPN significantly improves over existing models, achieving up to 2.2 improvement in detection accuracy (e.g., 36.3 for FPN vs. 38.5 for TFPN on COCO Val-17). Our method can obtain the same accuracy as FPN with ResNet-101 based on ResNet-50 and needs fewer parameters.
特征金字塔网络(Feature Pyramid Networks,简称FPN)是目前最流行的目标检测网络之一,它可以通过增强浅层特征来改善小目标的检测。然而,通过更深层次的特征增强来改进大目标检测的研究却很少。现有的一种方法是将不同层的特征图合并成一个新的特征图用于目标检测,但这可能导致噪声增加和信息丢失。另一种方法是在FPN的特征金字塔之后增加一个自下而上的结构,将浅层信息叠加到深层特征图中,但削弱了FPN检测小目标的强度。为了解决这些挑战,本文提出了TFPN(双特征金字塔网络),它包括:(1)FPN+,一种自下而上的结构,可以提高大型目标的检测;(2) TPS,双金字塔结构,提高介质目标检测;(3)二者与FPN的创新融合,在保持FPN在小目标检测中的优势的同时,显著提高了大中型目标的检测精度。使用MSCOCO目标检测数据集和BDD100K自动驾驶数据集进行的大量实验表明,TFPN比现有模型有了显著改善,检测精度提高了2.2(例如,FPN的36.3比COCO var -17上的TFPN的38.5)。该方法可以获得与基于ResNet-50的ResNet-101的FPN相同的精度,并且需要更少的参数。
{"title":"TFPN: Twin Feature Pyramid Networks for Object Detection","authors":"Yi Liang, Changjian Wang, Fangzhao Li, Yuxing Peng, Q. Lv, Yuan Yuan, Zhen Huang","doi":"10.1109/ICTAI.2019.00251","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00251","url":null,"abstract":"FPN (Feature Pyramid Networks) is one of the most popular object detection networks, which can improve small object detection by enhancing shallow features. However, limited attention has been paid to the improvement of large object detection via deeper feature enhancement. One existing approach merges the feature maps of different layers into a new feature map for object detection, but can lead to increased noise and loss of information. The other approach adds a bottom-up structure after the feature pyramid of FPN, which superimposes the information from shallow layers into the deep feature map but weakens the strength of FPN in detecting small objects. To address these challenges, this paper proposes TFPN (Twin Feature Pyramid Networks), which consists of (1) FPN+, a bottom-up structure that improves large object detection; (2) TPS, a Twin Pyramid Structure that improves medium object detection; and (3) innovative integration of these two with FPN, which can significantly improve the detection accuracy of large and medium objects while maintaining the advantage of FPN in small object detection. Extensive experiments using the MSCOCO object detection datasets and the BDD100K automatic driving dataset demonstrate that TFPN significantly improves over existing models, achieving up to 2.2 improvement in detection accuracy (e.g., 36.3 for FPN vs. 38.5 for TFPN on COCO Val-17). Our method can obtain the same accuracy as FPN with ResNet-101 based on ResNet-50 and needs fewer parameters.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132402450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
The Application of Network Based Embedding in Local Topic Detection from Social Media 基于网络的嵌入在社交媒体局部话题检测中的应用
Junsha Chen, Neng Gao, Cong Xue, Yifei Zhang, Chenyang Tu, Min Li
Detecting local topic from social media is an important task for many applications, such as local event discovery and activity recommendation. Recent years have witnessed growing interest in utilizing spatio-temporal social media for local topic detection. However, conventional topic models consider keywords as independent items, which suffer great limitations in modeling short texts from social media. Therefore, some studies introduce embedding into topic models to preserve the semantic correlation among keywords of short texts. Nevertheless, due to the lack of rich contexts in social media, the performance of these embedding based topic models still remain unsatisfactory. In order to enrich the contexts of keywords, we propose two network based embedding methods, both of which can generate rich contexts for keywords by random walks and produce coherent keyword embeddings for topic modeling. Besides, processing continuous spatio-temporal information in social media is also very challenging. Most of the existing methods simply split time and location into equal-size units, which fall short in capturing the continuity of spatio-temporal information. To address this issue, we present a hotspot detection algorithm to identify spatial and temporal hotspots, which can address spatio-temporal continuity and alleviate data sparsity. Finally, the experiments show that the performance of our methods has been improved significantly compared to the state-of-the-art methods.
从社交媒体中检测本地话题是许多应用程序的重要任务,例如本地事件发现和活动推荐。近年来,人们对利用时空社交媒体进行局部话题检测的兴趣日益浓厚。然而,传统的主题模型将关键词视为独立的项目,这在对社交媒体短文本建模时存在很大的局限性。因此,一些研究在主题模型中引入嵌入,以保持短文本关键词之间的语义相关性。然而,由于社交媒体中缺乏丰富的上下文,这些基于嵌入的主题模型的性能仍然令人不满意。为了丰富关键词上下文,我们提出了两种基于网络的嵌入方法,这两种方法都可以通过随机漫步生成丰富的关键词上下文,并产生连贯的关键词嵌入用于主题建模。此外,在社交媒体中处理连续的时空信息也非常具有挑战性。现有的方法大多是简单地将时间和位置分割成大小相等的单元,在捕捉时空信息的连续性方面存在不足。为了解决这一问题,我们提出了一种热点检测算法来识别时空热点,从而解决了时空连续性问题,减轻了数据的稀疏性。最后,实验表明,与最先进的方法相比,我们的方法的性能有了显着提高。
{"title":"The Application of Network Based Embedding in Local Topic Detection from Social Media","authors":"Junsha Chen, Neng Gao, Cong Xue, Yifei Zhang, Chenyang Tu, Min Li","doi":"10.1109/ICTAI.2019.00182","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00182","url":null,"abstract":"Detecting local topic from social media is an important task for many applications, such as local event discovery and activity recommendation. Recent years have witnessed growing interest in utilizing spatio-temporal social media for local topic detection. However, conventional topic models consider keywords as independent items, which suffer great limitations in modeling short texts from social media. Therefore, some studies introduce embedding into topic models to preserve the semantic correlation among keywords of short texts. Nevertheless, due to the lack of rich contexts in social media, the performance of these embedding based topic models still remain unsatisfactory. In order to enrich the contexts of keywords, we propose two network based embedding methods, both of which can generate rich contexts for keywords by random walks and produce coherent keyword embeddings for topic modeling. Besides, processing continuous spatio-temporal information in social media is also very challenging. Most of the existing methods simply split time and location into equal-size units, which fall short in capturing the continuity of spatio-temporal information. To address this issue, we present a hotspot detection algorithm to identify spatial and temporal hotspots, which can address spatio-temporal continuity and alleviate data sparsity. Finally, the experiments show that the performance of our methods has been improved significantly compared to the state-of-the-art methods.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131539046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Exploring Numerical Calculations with CalcNet 探索数值计算与CalcNet
Ashish Rana, A. Malhi, Kary Främling
Neural networks are not great generalizers outside their training range i.e. they are good at capturing bias but might miss the overall concept. Important issues with neural networks is that when testing data goes outside training range they fail to predict accurate results. Hence, they loose the ability to generalize a concept. For systematic numeric exploration neural accumulators (NAC) and neural arithmetic logic unit(NALU) are proposed which performs excellent for simple arithmetic operations. But, major limitation with these units is that they can't handle complex mathematical operations & equations. For example, NALU can predict accurate results for multiplication operation but not for factorial function which is essentially composition of multiplication operations only. It is unable to comprehend pattern behind an expression when composition of operations are involved. Hence, we propose a new neural network structure effectively which takes in complex compositional mathematical operations and produces best possible results with small NALU based neural networks as its pluggable modules which evaluates these expression at unitary level in a bottom-up manner. We call this effective neural network as CalcNet, as it helps in predicting accurate calculations for complex numerical expressions even for values that are out of training range. As part of our study we applied this network on numerically approximating complex equations, evaluating biquadratic equations and tested reusability of these modules. We arrived at far better generalizations for complex arithmetic extrapolation tasks as compare to both only NALU layer based neural networks and simple feed forward neural networks. Also, we achieved even better results for our golden ratio based modified NAC and NALU structures for both interpolating and extrapolating tasks in all evaluation experiments. Finally, from reusability standpoint this model demonstrate strong invariance for making predictions on different tasks.
神经网络在其训练范围之外并不是很好的泛化者,也就是说,它们善于捕捉偏见,但可能会错过整体概念。神经网络的一个重要问题是,当测试数据超出训练范围时,它们无法预测准确的结果。因此,他们失去了概括一个概念的能力。针对系统的数值探索,提出了神经累加器(NAC)和神经算术逻辑单元(NALU),它们能很好地处理简单的算术运算。但是,这些单元的主要限制是它们不能处理复杂的数学运算和方程。例如,NALU可以预测乘法操作的准确结果,但不能预测阶乘函数,因为阶乘函数本质上只是乘法操作的组合。当涉及到操作组合时,无法理解表达式背后的模式。因此,我们提出了一种新的神经网络结构,它有效地处理复杂的组合数学运算,并以基于NALU的小型神经网络作为其可插拔模块,以自下而上的方式在酉级上评估这些表达式,从而产生最佳结果。我们称这种有效的神经网络为CalcNet,因为它有助于预测复杂数值表达式的精确计算,甚至是超出训练范围的值。作为我们研究的一部分,我们将该网络应用于数值逼近复杂方程,评估双二次方程并测试这些模块的可重用性。与仅基于NALU层的神经网络和简单的前馈神经网络相比,我们在复杂的算术外推任务中得到了更好的泛化。此外,在所有评估实验中,我们基于黄金分割的改进NAC和NALU结构在内插和外推任务中都取得了更好的结果。最后,从可重用性的角度来看,该模型在对不同任务进行预测时表现出很强的不变性。
{"title":"Exploring Numerical Calculations with CalcNet","authors":"Ashish Rana, A. Malhi, Kary Främling","doi":"10.1109/ICTAI.2019.00192","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00192","url":null,"abstract":"Neural networks are not great generalizers outside their training range i.e. they are good at capturing bias but might miss the overall concept. Important issues with neural networks is that when testing data goes outside training range they fail to predict accurate results. Hence, they loose the ability to generalize a concept. For systematic numeric exploration neural accumulators (NAC) and neural arithmetic logic unit(NALU) are proposed which performs excellent for simple arithmetic operations. But, major limitation with these units is that they can't handle complex mathematical operations & equations. For example, NALU can predict accurate results for multiplication operation but not for factorial function which is essentially composition of multiplication operations only. It is unable to comprehend pattern behind an expression when composition of operations are involved. Hence, we propose a new neural network structure effectively which takes in complex compositional mathematical operations and produces best possible results with small NALU based neural networks as its pluggable modules which evaluates these expression at unitary level in a bottom-up manner. We call this effective neural network as CalcNet, as it helps in predicting accurate calculations for complex numerical expressions even for values that are out of training range. As part of our study we applied this network on numerically approximating complex equations, evaluating biquadratic equations and tested reusability of these modules. We arrived at far better generalizations for complex arithmetic extrapolation tasks as compare to both only NALU layer based neural networks and simple feed forward neural networks. Also, we achieved even better results for our golden ratio based modified NAC and NALU structures for both interpolating and extrapolating tasks in all evaluation experiments. Finally, from reusability standpoint this model demonstrate strong invariance for making predictions on different tasks.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117279029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Shallow Deep Learning: Embedding Verbatim K-Means in Deep Neural Networks 浅深度学习:在深度神经网络中逐字嵌入k均值
Len Du
In this paper we show how to implement a deep neural network that is strictly equivalent (sans floating-point errors) to the verbatim (batch) k-means algorithm or Lloyd's algorithm, when trained with gradient descent. Most interestingly, doing so shows that the k-means algorithm, a staple of "conventional'" or "shallow'" machine learning, can actually be seen as a special case of deep learning, contrary to the general perception that deep learning is a subset of machine learning. Doing so also automatically introduces yet another unsupervised learning technique into the arsenal of deep learning, which happens to be an example of interpretable deep neural networks as well. Finally, we also show how to utilize the powerful deep learning infrastructures with very little extra effort for adaptation.
在本文中,我们展示了如何实现一个深度神经网络,当使用梯度下降训练时,它与逐字(批处理)k-均值算法或劳埃德算法严格等效(无浮点误差)。最有趣的是,这样做表明,作为“传统”或“浅”机器学习的主要内容,k-means算法实际上可以被视为深度学习的一个特例,这与深度学习是机器学习的一个子集的普遍看法相反。这样做也会自动将另一种无监督学习技术引入深度学习的武器库,这恰好也是可解释深度神经网络的一个例子。最后,我们还展示了如何利用强大的深度学习基础设施,而无需额外的适应工作。
{"title":"Shallow Deep Learning: Embedding Verbatim K-Means in Deep Neural Networks","authors":"Len Du","doi":"10.1109/ICTAI.2019.00035","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00035","url":null,"abstract":"In this paper we show how to implement a deep neural network that is strictly equivalent (sans floating-point errors) to the verbatim (batch) k-means algorithm or Lloyd's algorithm, when trained with gradient descent. Most interestingly, doing so shows that the k-means algorithm, a staple of \"conventional'\" or \"shallow'\" machine learning, can actually be seen as a special case of deep learning, contrary to the general perception that deep learning is a subset of machine learning. Doing so also automatically introduces yet another unsupervised learning technique into the arsenal of deep learning, which happens to be an example of interpretable deep neural networks as well. Finally, we also show how to utilize the powerful deep learning infrastructures with very little extra effort for adaptation.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114931452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Semi-Supervised Cross-Modal Hashing Based on Label Prediction and Distance Preserving 基于标签预测和距离保持的半监督跨模态哈希
Xu Zhang, Xin Tian, Bing Yang, Zuyu Zhang, Yan Li
Unlabeled data can be easily collected and help to exploit the correlations among different modalities. Existing works tried to explore label information contained in unlabeled data, however most of them suffer from difficulties in separating samples from different categories and have great interference. This paper proposes a novel method named semi-supervised cross-modal hashing based on label prediction and distance preserving(SS-LPDP). First, we use the deep neural networks to extract the feature of the labeled data among different modalities and get the feature distribution of each category. Second, the similarity of the data among different modalities is maximized based on the extracted feature and the label information. A common objective function is proposed with distance preserving constraint, which can effectively separate data into different categories and reduce interference in retrieval. An optimization algorithm is used to update the network parameters of feature learning in each modality, and the label information of unlabeled data are dynamically updated according to the changes of the feature distribution in each iteration. Experimental evaluation on Wiki, Pascal and NUS-WIDE datasets show that the proposed method outperforms recent methods when we set 25% samples without category labels.
未标记的数据可以很容易地收集,并有助于开发不同模式之间的相关性。现有的工作试图挖掘未标记数据中包含的标签信息,但大多数工作都存在从不同类别中分离样本的困难和很大的干扰。提出了一种基于标签预测和距离保持的半监督跨模态哈希算法(SS-LPDP)。首先,利用深度神经网络提取标记数据在不同模态间的特征,得到各类别的特征分布;其次,基于提取的特征和标签信息,最大化不同模态间数据的相似度;提出了一种带距离保持约束的公共目标函数,可以有效地将数据分类,减少检索过程中的干扰。采用优化算法更新各模态特征学习的网络参数,并根据每次迭代中特征分布的变化动态更新未标记数据的标签信息。在Wiki、Pascal和NUS-WIDE数据集上的实验评估表明,当我们设置25%的样本不带类别标签时,所提出的方法优于现有的方法。
{"title":"Semi-Supervised Cross-Modal Hashing Based on Label Prediction and Distance Preserving","authors":"Xu Zhang, Xin Tian, Bing Yang, Zuyu Zhang, Yan Li","doi":"10.1109/ICTAI.2019.00053","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00053","url":null,"abstract":"Unlabeled data can be easily collected and help to exploit the correlations among different modalities. Existing works tried to explore label information contained in unlabeled data, however most of them suffer from difficulties in separating samples from different categories and have great interference. This paper proposes a novel method named semi-supervised cross-modal hashing based on label prediction and distance preserving(SS-LPDP). First, we use the deep neural networks to extract the feature of the labeled data among different modalities and get the feature distribution of each category. Second, the similarity of the data among different modalities is maximized based on the extracted feature and the label information. A common objective function is proposed with distance preserving constraint, which can effectively separate data into different categories and reduce interference in retrieval. An optimization algorithm is used to update the network parameters of feature learning in each modality, and the label information of unlabeled data are dynamically updated according to the changes of the feature distribution in each iteration. Experimental evaluation on Wiki, Pascal and NUS-WIDE datasets show that the proposed method outperforms recent methods when we set 25% samples without category labels.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"276 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114945307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Harnessing GAN with Metric Learning for One-Shot Generation on a Fine-Grained Category 基于度量学习的GAN在细粒度分类上的一次性生成
Yusuke Ohtsubo, Tetsu Matsukawa, Einoshin Suzuki
We propose a GAN-based one-shot generation method on a fine-grained category, which represents a subclass of a category, typically with diverse examples. One-shot generation refers to a task of taking an image which belongs to a class not used in the training phase and then generating a set of new images belonging to the same class. Generative Adversarial Network (GAN), which represents a type of deep neural networks with competing generator and discriminator, has proven to be useful in generating realistic images. Especially DAGAN, which maps the input image to a low-dimensional space via an encoder and then back to the example space via a decoder, has been quite effective with datasets such as handwritten character datasets. However, when the class corresponds to a fine-grained category, DAGAN occasionally generates images which are regarded as belonging to other classes due to the rich variety of the examples in the class and the low dissimilarities of the examples among the classes. For example, it accidentally generates facial images of different persons when the class corresponds to a specific person. To circumvent this problem, we introduce a metric learning with a triplet loss to the bottleneck layer of DAGAN to penalize such a generation. We also extend the optimization algorithm of DAGAN to an alternating procedure for two types of loss functions. Our proposed method outperforms DAGAN in the GAN-test task for VGG-Face dataset and CompCars dataset by 5.6% and 4.8% in accuracy, respectively. We also conducted experiments for the data augmentation task and observed 4.5% higher accuracy for our proposed method over DAGAN for VGG-Face dataset.
我们提出了一种基于gan的细粒度类别的一次性生成方法,细粒度类别代表一个类别的子类,通常具有不同的示例。一次性生成(One-shot generation)是指取一张训练阶段未使用的类别的图像,然后生成一组属于同一类别的新图像。生成式对抗网络(GAN)是一种具有生成器和鉴别器竞争的深度神经网络,在生成逼真图像方面非常有用。特别是DAGAN,它通过编码器将输入图像映射到低维空间,然后通过解码器返回到示例空间,对于手写字符数据集等数据集非常有效。然而,当类对应于一个细粒度的类别时,DAGAN偶尔会产生被认为属于其他类的图像,因为类中的样本种类丰富,类之间的样本不相似度很低。例如,当类对应于特定的人时,它会意外地生成不同人的面部图像。为了避免这个问题,我们在DAGAN的瓶颈层引入了一个带有三重损失的度量学习来惩罚这样的生成。我们还将DAGAN的优化算法推广到两类损失函数的交替过程。在VGG-Face数据集和CompCars数据集的gan测试任务中,我们提出的方法的准确率分别比DAGAN高5.6%和4.8%。我们还对VGG-Face数据集的数据增强任务进行了实验,发现我们提出的方法在DAGAN上的准确率提高了4.5%。
{"title":"Harnessing GAN with Metric Learning for One-Shot Generation on a Fine-Grained Category","authors":"Yusuke Ohtsubo, Tetsu Matsukawa, Einoshin Suzuki","doi":"10.1109/ICTAI.2019.00130","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00130","url":null,"abstract":"We propose a GAN-based one-shot generation method on a fine-grained category, which represents a subclass of a category, typically with diverse examples. One-shot generation refers to a task of taking an image which belongs to a class not used in the training phase and then generating a set of new images belonging to the same class. Generative Adversarial Network (GAN), which represents a type of deep neural networks with competing generator and discriminator, has proven to be useful in generating realistic images. Especially DAGAN, which maps the input image to a low-dimensional space via an encoder and then back to the example space via a decoder, has been quite effective with datasets such as handwritten character datasets. However, when the class corresponds to a fine-grained category, DAGAN occasionally generates images which are regarded as belonging to other classes due to the rich variety of the examples in the class and the low dissimilarities of the examples among the classes. For example, it accidentally generates facial images of different persons when the class corresponds to a specific person. To circumvent this problem, we introduce a metric learning with a triplet loss to the bottleneck layer of DAGAN to penalize such a generation. We also extend the optimization algorithm of DAGAN to an alternating procedure for two types of loss functions. Our proposed method outperforms DAGAN in the GAN-test task for VGG-Face dataset and CompCars dataset by 5.6% and 4.8% in accuracy, respectively. We also conducted experiments for the data augmentation task and observed 4.5% higher accuracy for our proposed method over DAGAN for VGG-Face dataset.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115619919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dynamic Multi-population Artificial Bee Colony Algorithm 动态多种群人工蜂群算法
Xinyu Zhou, Yiwen Ling, M. Zhong, Mingwen Wang
As a relatively new paradigm of bio-inspired optimization techniques, artificial bee colony (ABC) algorithm has attracted much attention for its simplicity and effectiveness. However, the performance of ABC is not satisfactory when solving some complex optimization problems. To improve its performance, we propose a novel ABC variant by designing a dynamic multi-population scheme (DMPS). In DMPS, the population is divided into several subpopulations, and the size of subpopulation is adjusted dynamically by checking the quality of the global best solution. Moreover, we design two novel solution search equations to maximize the effectiveness of DMPS, in which the local best solution of each subpopulation and the global best solution of the whole population are utilized simultaneously. In the experiments, 32 widely used benchmark functions are used, and four well-established ABC variants are involved in the comparison. The comparative results show that our approach performs better on the majority of benchmark functions.
人工蜂群算法(artificial bee colony, ABC)作为一种较新的仿生优化技术,因其简单、有效而备受关注。然而,在解决一些复杂的优化问题时,ABC算法的性能并不令人满意。为了提高ABC算法的性能,我们通过设计动态多种群方案(DMPS)提出了一种新的ABC变体。在DMPS中,将种群划分为若干个子种群,并通过检查全局最优解的质量来动态调整子种群的大小。此外,为了使DMPS的有效性最大化,我们设计了两个新的解搜索方程,其中每个子种群的局部最优解和整个种群的全局最优解同时被利用。在实验中,使用了32个广泛使用的基准函数,并涉及4个完善的ABC变体进行比较。对比结果表明,我们的方法在大多数基准函数上表现更好。
{"title":"Dynamic Multi-population Artificial Bee Colony Algorithm","authors":"Xinyu Zhou, Yiwen Ling, M. Zhong, Mingwen Wang","doi":"10.1109/ICTAI.2019.00113","DOIUrl":"https://doi.org/10.1109/ICTAI.2019.00113","url":null,"abstract":"As a relatively new paradigm of bio-inspired optimization techniques, artificial bee colony (ABC) algorithm has attracted much attention for its simplicity and effectiveness. However, the performance of ABC is not satisfactory when solving some complex optimization problems. To improve its performance, we propose a novel ABC variant by designing a dynamic multi-population scheme (DMPS). In DMPS, the population is divided into several subpopulations, and the size of subpopulation is adjusted dynamically by checking the quality of the global best solution. Moreover, we design two novel solution search equations to maximize the effectiveness of DMPS, in which the local best solution of each subpopulation and the global best solution of the whole population are utilized simultaneously. In the experiments, 32 widely used benchmark functions are used, and four well-established ABC variants are involved in the comparison. The comparative results show that our approach performs better on the majority of benchmark functions.","PeriodicalId":346657,"journal":{"name":"2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116179845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2019 IEEE 31st International Conference on Tools with Artificial Intelligence (ICTAI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1