首页 > 最新文献

2022 14th International Conference on Knowledge and Smart Technology (KST)最新文献

英文 中文
Comparison Analysis of Data Augmentation using Bootstrap, GANs and Autoencoder 自举、gan和自编码器数据增强的比较分析
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9729065
Mukrin Nakhwan, Rakkrit Duangsoithong
In order to improve predictive accuracy for insufficient observations, data augmentation is a well-known and commonly useful technique to increase more samples by generating new data which can avoid data collection problems. This paper presents comparison analysis of three data augmentation methods using Bootstrap method, Generative Adversarial Networks (GANs) and Autoencoder for increasing a number of samples. The proposal is applied on 8 datasets with binary classification from repository data websites. The research is mainly evaluated by generating new additional data using data augmentation. Secondly, combining generated samples and original data. Finally, validating performance on four classifier models. The experimental result showed that the proposed approach of increasing samples by Autoencoder and GANs achieved better predictive performance than the original data. Conversely, increasing samples by Bootstrap method provided lowest predictive performance.
为了在观测不足的情况下提高预测精度,数据扩增是一种众所周知且常用的技术,通过生成新数据来增加更多的样本,从而避免数据收集问题。本文对自举法、生成对抗网络(GANs)和自动编码器三种数据增强方法进行了比较分析。将该方法应用于知识库数据网站的8个二元分类数据集。该研究主要通过数据增强产生新的附加数据来评估。其次,将生成的样本与原始数据相结合。最后,验证了四种分类器模型的性能。实验结果表明,采用自编码器和gan增加样本的方法比原始数据具有更好的预测性能。相反,通过Bootstrap方法增加样本的预测性能最低。
{"title":"Comparison Analysis of Data Augmentation using Bootstrap, GANs and Autoencoder","authors":"Mukrin Nakhwan, Rakkrit Duangsoithong","doi":"10.1109/KST53302.2022.9729065","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9729065","url":null,"abstract":"In order to improve predictive accuracy for insufficient observations, data augmentation is a well-known and commonly useful technique to increase more samples by generating new data which can avoid data collection problems. This paper presents comparison analysis of three data augmentation methods using Bootstrap method, Generative Adversarial Networks (GANs) and Autoencoder for increasing a number of samples. The proposal is applied on 8 datasets with binary classification from repository data websites. The research is mainly evaluated by generating new additional data using data augmentation. Secondly, combining generated samples and original data. Finally, validating performance on four classifier models. The experimental result showed that the proposed approach of increasing samples by Autoencoder and GANs achieved better predictive performance than the original data. Conversely, increasing samples by Bootstrap method provided lowest predictive performance.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125194690","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Traffic Light and Crosswalk Detection and Localization Using Vehicular Camera 基于车载摄像头的红绿灯和人行横道检测与定位
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9729066
S. Wangsiripitak, Keisuke Hano, S. Kuchii
An improved convolutional neural network model for traffic light and crosswalk detection and localization using visual information from a vehicular camera is proposed. Yolov4 darknet and its pretrained model are used in transfer learning using our datasets of traffic lights and crosswalks; the trained model is supposed to be used for red-light running detection of the preceding vehicle. Experimental results, compared to the result of the pretrained model learned only from the Microsoft COCO dataset, showed an improved performance of traffic light detection on our test images which were taken under various lighting conditions and interferences; 36.91% higher recall and 39.21% less false positive rate. The crosswalk, which is incapable of detection in the COCO model, could be detected with 93.37% recall and 7.74% false-positive rate.
提出了一种改进的卷积神经网络模型,利用车载摄像头的视觉信息对红绿灯和人行横道进行检测和定位。Yolov4 darknet及其预训练模型用于迁移学习,使用我们的交通灯和人行横道数据集;训练后的模型用于检测前车的闯红灯。实验结果表明,与仅从Microsoft COCO数据集学习的预训练模型的结果相比,我们在各种光照条件和干扰下拍摄的测试图像上的交通灯检测性能有所提高;召回率提高36.91%,假阳性率降低39.21%。在COCO模型中无法检测到的人行横道,其召回率为93.37%,假阳性率为7.74%。
{"title":"Traffic Light and Crosswalk Detection and Localization Using Vehicular Camera","authors":"S. Wangsiripitak, Keisuke Hano, S. Kuchii","doi":"10.1109/KST53302.2022.9729066","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9729066","url":null,"abstract":"An improved convolutional neural network model for traffic light and crosswalk detection and localization using visual information from a vehicular camera is proposed. Yolov4 darknet and its pretrained model are used in transfer learning using our datasets of traffic lights and crosswalks; the trained model is supposed to be used for red-light running detection of the preceding vehicle. Experimental results, compared to the result of the pretrained model learned only from the Microsoft COCO dataset, showed an improved performance of traffic light detection on our test images which were taken under various lighting conditions and interferences; 36.91% higher recall and 39.21% less false positive rate. The crosswalk, which is incapable of detection in the COCO model, could be detected with 93.37% recall and 7.74% false-positive rate.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"99 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126660901","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Developing an Automatic Speech Recognizer For Filipino with English Code-Switching in News Broadcast 新闻广播中英语语码转换的菲律宾语自动语音识别器的研制
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9727235
Mark Louis Lim, A. J. Xu, C. Lin, Zi-He Chen, Ronald M. Pascual
Closed-captioning systems are well-known for video-based broadcasting companies as society transitions into internet-based information consumption. These captioning systems are utilized to cater to most consumers. However, a captioning system for the Filipino language is not readily available to the public. News anchors in the Philippines tend to incorporate a code-switching behavior that mixes English and Filipino languages, which are the two major languages that Filipinos use. The goal of this research is to develop an automatic speech recognizer (ASR) for a captioning system for Filipino news broadcast domain videos. Experiments on finding the optimal speech models and features, and on how code-switching affects the system were conducted. Best results were obtained by using linear discriminant analysis with maximum likelihood linear transform (LDA+MLLT) and speaker adaptive training (SAT) for acoustic modeling. Initial investigation also shows that there is no general pattern for the ASR's performance as a function of code-switching frequency.
随着社会向基于互联网的信息消费转型,封闭式字幕系统在以视频为基础的广播公司中非常有名。这些字幕系统被用来迎合大多数消费者。然而,菲律宾语的字幕系统还没有提供给公众。菲律宾的新闻主播倾向于结合英语和菲律宾语的代码转换行为,这是菲律宾人使用的两种主要语言。本研究的目的是为菲律宾新闻广播领域视频的字幕系统开发一个自动语音识别器(ASR)。进行了寻找最优语音模型和特征的实验,以及语码转换对系统的影响。采用最大似然线性变换线性判别分析(LDA+MLLT)和说话人自适应训练(SAT)进行声学建模,效果最好。初步调查还表明,ASR的性能作为代码转换频率的函数没有一般模式。
{"title":"Developing an Automatic Speech Recognizer For Filipino with English Code-Switching in News Broadcast","authors":"Mark Louis Lim, A. J. Xu, C. Lin, Zi-He Chen, Ronald M. Pascual","doi":"10.1109/KST53302.2022.9727235","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9727235","url":null,"abstract":"Closed-captioning systems are well-known for video-based broadcasting companies as society transitions into internet-based information consumption. These captioning systems are utilized to cater to most consumers. However, a captioning system for the Filipino language is not readily available to the public. News anchors in the Philippines tend to incorporate a code-switching behavior that mixes English and Filipino languages, which are the two major languages that Filipinos use. The goal of this research is to develop an automatic speech recognizer (ASR) for a captioning system for Filipino news broadcast domain videos. Experiments on finding the optimal speech models and features, and on how code-switching affects the system were conducted. Best results were obtained by using linear discriminant analysis with maximum likelihood linear transform (LDA+MLLT) and speaker adaptive training (SAT) for acoustic modeling. Initial investigation also shows that there is no general pattern for the ASR's performance as a function of code-switching frequency.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125918631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Development of Anomaly Detection Model for Welding Classification Using Arc Sound 基于电弧声的焊接分类异常检测模型的开发
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9729058
Phongsin Jirapipattanaporn, Worawat Lawanont
This study introduces the method to classify weld bead type from arc sound of the gas metal arc welding process by applying machine learning techniques. In this research, we mainly focused on two types of weld bead which were normal weld bead and burn-through weld bead. The signal processing technique was implemented in this work to visualize welding sound data, recorded with a microphone array. All recorded sounds are imported for generating the spectrogram using Python programming and Fourier transformation to analyze and explore the difference of each sound that occurred from different weld bead types. The feature extraction from the sound data is used to construct the dataset for developing the model. Three machine learning models were trained by three different algorithms. Which were recurrent neural network (RNN), Long-short Term Memory (LSTM), and one-class Support Vector Machine (one-class SVM). Each model was evaluated with accuracy and confusion matrix. After a train and testing each model, the result showed that each model performs with an overall accuracy greater than 80 percent for each model. Given the performance of the model developed in this research, these models can be applied to the welding process. And the method from this research can also be applied with another manufacturing process in future work.
介绍了利用机器学习技术从气体金属弧焊过程的电弧声中识别焊缝类型的方法。在本研究中,我们主要研究了两种类型的焊头,即普通焊头和烧透焊头。本文采用信号处理技术对焊接声数据进行可视化处理,并通过麦克风阵列进行记录。输入所有录制的声音,使用Python编程和傅立叶变换生成频谱图,分析和探索不同焊头类型产生的每种声音的差异。从声音数据中提取特征用于构建数据集,用于开发模型。三个机器学习模型通过三种不同的算法进行训练。分别是递归神经网络(RNN)、长短期记忆(LSTM)和一类支持向量机(SVM)。用准确度和混淆矩阵对每个模型进行评价。在对每个模型进行训练和测试后,结果表明每个模型的总体准确率都大于80%。考虑到本研究建立的模型的性能,这些模型可以应用于焊接过程。本研究的方法也可以应用到其他制造工艺中。
{"title":"Development of Anomaly Detection Model for Welding Classification Using Arc Sound","authors":"Phongsin Jirapipattanaporn, Worawat Lawanont","doi":"10.1109/KST53302.2022.9729058","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9729058","url":null,"abstract":"This study introduces the method to classify weld bead type from arc sound of the gas metal arc welding process by applying machine learning techniques. In this research, we mainly focused on two types of weld bead which were normal weld bead and burn-through weld bead. The signal processing technique was implemented in this work to visualize welding sound data, recorded with a microphone array. All recorded sounds are imported for generating the spectrogram using Python programming and Fourier transformation to analyze and explore the difference of each sound that occurred from different weld bead types. The feature extraction from the sound data is used to construct the dataset for developing the model. Three machine learning models were trained by three different algorithms. Which were recurrent neural network (RNN), Long-short Term Memory (LSTM), and one-class Support Vector Machine (one-class SVM). Each model was evaluated with accuracy and confusion matrix. After a train and testing each model, the result showed that each model performs with an overall accuracy greater than 80 percent for each model. Given the performance of the model developed in this research, these models can be applied to the welding process. And the method from this research can also be applied with another manufacturing process in future work.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130464246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Novel Relational Deep Network for Single Object Tracking 一种用于单目标跟踪的新型关系深度网络
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9729070
Pimpa Cheewaprakobkit, T. Shih, Chih-Yang Lin, Hung-Chun Liao
Virtual object tracking is an active research area in computer vision. It aims to estimate the location of the target object in video frames. For the past few years, the deep learning method has been widely used for object tracking to improve accuracy. However, there are still challenges of performance problems and accuracy. This study aims to enhance the performance of an object detection model by focusing on single object tracking using Siamese network architecture and a correlation filter to find the relationship between the target object and search object from a series of continuous images. We mitigate some challenging problems in the Siamese network by adding variance loss to improve the model to distinguish between the foreground and the background. Furthermore, we add the attention mechanism and process the cropped image to find the relationship between objects and objects. Our experiment used the VOT2019 dataset for testing object tracking and the CUHK03 dataset for the training model. The result demonstrates that the proposed model achieves promising prediction performance to solve the image occlusion problem and reduce false alarms from object detection. We achieved an accuracy of 0.608, a robustness of 0.539, and an expected average overlap (EAO) score of 0.217. Our tracker runs at approximately 26 fps on GPU.
虚拟目标跟踪是计算机视觉中一个活跃的研究领域。它的目的是估计视频帧中目标物体的位置。在过去的几年里,深度学习方法被广泛用于目标跟踪,以提高准确性。然而,仍然存在性能问题和准确性方面的挑战。本研究旨在通过使用Siamese网络架构和相关滤波器从一系列连续图像中寻找目标对象与搜索对象之间的关系,重点关注单个目标跟踪,从而提高目标检测模型的性能。我们通过加入方差损失来改善Siamese网络中一些具有挑战性的问题,以区分前景和背景。在此基础上,我们增加了注意机制,并对裁剪后的图像进行处理,寻找物体与物体之间的关系。我们的实验使用VOT2019数据集来测试目标跟踪,使用CUHK03数据集来训练模型。结果表明,该模型在解决图像遮挡问题和减少目标检测虚警方面取得了较好的预测效果。我们获得了0.608的准确性,0.539的鲁棒性,以及0.217的预期平均重叠(EAO)得分。我们的跟踪器在GPU上以大约26 fps的速度运行。
{"title":"A Novel Relational Deep Network for Single Object Tracking","authors":"Pimpa Cheewaprakobkit, T. Shih, Chih-Yang Lin, Hung-Chun Liao","doi":"10.1109/KST53302.2022.9729070","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9729070","url":null,"abstract":"Virtual object tracking is an active research area in computer vision. It aims to estimate the location of the target object in video frames. For the past few years, the deep learning method has been widely used for object tracking to improve accuracy. However, there are still challenges of performance problems and accuracy. This study aims to enhance the performance of an object detection model by focusing on single object tracking using Siamese network architecture and a correlation filter to find the relationship between the target object and search object from a series of continuous images. We mitigate some challenging problems in the Siamese network by adding variance loss to improve the model to distinguish between the foreground and the background. Furthermore, we add the attention mechanism and process the cropped image to find the relationship between objects and objects. Our experiment used the VOT2019 dataset for testing object tracking and the CUHK03 dataset for the training model. The result demonstrates that the proposed model achieves promising prediction performance to solve the image occlusion problem and reduce false alarms from object detection. We achieved an accuracy of 0.608, a robustness of 0.539, and an expected average overlap (EAO) score of 0.217. Our tracker runs at approximately 26 fps on GPU.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129671759","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VAPE-BRIDGE: Bridging OpenVAS Results for Automating Metasploit Framework VAPE-BRIDGE:桥接OpenVAS结果用于自动化Metasploit框架
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9729085
Kankanok Vimala, S. Fugkeaw
Vulnerability assessment (VA) and penetration test (PenTest) are required by many organizations to satisfy their security auditing and compliance. VA and PenTest are conducted in the different stage and they are done through the software tools. Implementing the system that is able to convert the VA scan result to be rendered in the PenTest tool is a real challenge. This paper proposes a design and development of a system called VAPE-BRIDGE that provides the automatic conversion of the scan result of Open Vulnerability assessment scanner (OpenVAS) to be the exploitable scripts that will be executed in the Metasploit which is a widely-used opensource PenTest program. Specifically, the tool is designed to automatically extract the vulnerabilities listed in Open Web Application Security Project 10 (OWASP 10) and exploit them to be tested in the Metasploit. Our VAPE-BRIDGE encompasses three main components including (1) Scan Result Extraction responsible for extracting the VA scan results related to OWASP10 (2) Target List Repository responsible for retaining lists of vulnerabilities to be used in the process of Metasploit, and (3) Automated Shell Scripts Exploitation responsible for generating the script to render the exploit module to be executed in Metasploit. For the implementation, the VAPE-Bridge protype system was tested with a number of test cases in converting the scan results into shell code and rendering results to be tested in Metasploit. The experimental results showed that the system is functionally correct for all cases.
许多组织需要漏洞评估(VA)和渗透测试(PenTest)来满足其安全审计和遵从性。VA和PenTest是在不同的阶段进行的,它们是通过软件工具完成的。实现能够将VA扫描结果转换为在PenTest工具中呈现的系统是一个真正的挑战。本文设计并开发了一个名为VAPE-BRIDGE的系统,该系统将开放漏洞评估扫描器(Open Vulnerability assessment scanner, OpenVAS)的扫描结果自动转换为可被利用的脚本,并在广泛使用的开源测试程序Metasploit中执行。具体来说,该工具旨在自动提取开放Web应用程序安全项目10 (OWASP 10)中列出的漏洞,并利用它们在Metasploit中进行测试。我们的VAPE-BRIDGE包含三个主要组件,包括(1)扫描结果提取,负责提取与OWASP10相关的VA扫描结果;(2)目标列表存储库,负责保留在Metasploit过程中使用的漏洞列表;(3)自动Shell脚本开发,负责生成脚本,以呈现在Metasploit中执行的漏洞模块。为了实现VAPE-Bridge原型系统,在将扫描结果转换为shell代码和渲染结果以在Metasploit中进行测试时,使用了许多测试用例对其进行了测试。实验结果表明,该系统在所有情况下都是功能正确的。
{"title":"VAPE-BRIDGE: Bridging OpenVAS Results for Automating Metasploit Framework","authors":"Kankanok Vimala, S. Fugkeaw","doi":"10.1109/KST53302.2022.9729085","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9729085","url":null,"abstract":"Vulnerability assessment (VA) and penetration test (PenTest) are required by many organizations to satisfy their security auditing and compliance. VA and PenTest are conducted in the different stage and they are done through the software tools. Implementing the system that is able to convert the VA scan result to be rendered in the PenTest tool is a real challenge. This paper proposes a design and development of a system called VAPE-BRIDGE that provides the automatic conversion of the scan result of Open Vulnerability assessment scanner (OpenVAS) to be the exploitable scripts that will be executed in the Metasploit which is a widely-used opensource PenTest program. Specifically, the tool is designed to automatically extract the vulnerabilities listed in Open Web Application Security Project 10 (OWASP 10) and exploit them to be tested in the Metasploit. Our VAPE-BRIDGE encompasses three main components including (1) Scan Result Extraction responsible for extracting the VA scan results related to OWASP10 (2) Target List Repository responsible for retaining lists of vulnerabilities to be used in the process of Metasploit, and (3) Automated Shell Scripts Exploitation responsible for generating the script to render the exploit module to be executed in Metasploit. For the implementation, the VAPE-Bridge protype system was tested with a number of test cases in converting the scan results into shell code and rendering results to be tested in Metasploit. The experimental results showed that the system is functionally correct for all cases.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129031018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Efficient Image Embedding for Fine-Grained Visual Classification 用于细粒度视觉分类的高效图像嵌入
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9729062
Soranan Payatsuporn, B. Kijsirikul
Fine-grained visual classification (FGVC) is a task belonging to multiple sub-categories classification. It is a challenging task due to high intraclass variation and inter-class similarity. Most exiting methods pay attention to capture discriminative semantic parts to address those problems. In this paper, we introduce a two-level network which consists of raw-level and object-level networks, and we name it “Efficient Image Embedding”. Its training procedure has two stages which the raw-level is for localization by the aggregation of feature maps, and the last is for classification. The two-level use Adaptive Angular Margin loss (AAM-loss), which improve an intra-class compactness and inter-class variety of image embedding. Our approach is to identify object regions without any hand-crafted bounding-box, and can be trained in an end-to-end manner. It has achieved better accuracy on two datasets compared to the existing work, which are 89.0% for CUB200-2011 and 93.3% for FGVC-Aircraft.
细粒度视觉分类(FGVC)是一种属于多个子类别的分类任务。由于类内差异和类间相似性较大,这是一项具有挑战性的任务。现有的大多数方法都注重捕获判别语义部分来解决这些问题。本文介绍了一种由原始级和对象级网络组成的两级网络,并将其命名为“高效图像嵌入”。其训练过程分为两个阶段,第一阶段是通过特征映射的聚合进行定位,第二阶段是进行分类。该方法采用自适应角边缘损失(AAM-loss),提高了类内图像嵌入的紧凑性和类间图像嵌入的多样性。我们的方法是在没有任何手工制作的边界盒的情况下识别对象区域,并且可以以端到端的方式进行训练。与现有工作相比,它在两个数据集上取得了更好的精度,其中CUB200-2011为89.0%,FGVC-Aircraft为93.3%。
{"title":"Efficient Image Embedding for Fine-Grained Visual Classification","authors":"Soranan Payatsuporn, B. Kijsirikul","doi":"10.1109/KST53302.2022.9729062","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9729062","url":null,"abstract":"Fine-grained visual classification (FGVC) is a task belonging to multiple sub-categories classification. It is a challenging task due to high intraclass variation and inter-class similarity. Most exiting methods pay attention to capture discriminative semantic parts to address those problems. In this paper, we introduce a two-level network which consists of raw-level and object-level networks, and we name it “Efficient Image Embedding”. Its training procedure has two stages which the raw-level is for localization by the aggregation of feature maps, and the last is for classification. The two-level use Adaptive Angular Margin loss (AAM-loss), which improve an intra-class compactness and inter-class variety of image embedding. Our approach is to identify object regions without any hand-crafted bounding-box, and can be trained in an end-to-end manner. It has achieved better accuracy on two datasets compared to the existing work, which are 89.0% for CUB200-2011 and 93.3% for FGVC-Aircraft.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"107 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122432574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Hybrid Deep Neural Network for Classifying Transportation Modes based on Human Activity Vibration 基于人体活动振动的混合深度神经网络交通方式分类
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9729079
S. Mekruksavanich, Ponnipa Jantawong, I. You, A. Jitpattanakul
Sensor advanced technologies have facilitated the growth of various solutions for recognizing human movement through wearable devices. Characterization of the means of transportation has become beneficial applications in an intelligent transportation system since it enables context-aware support for the implementation of systems such as driver assistance and intelligent transportation management. Smartphone sensing technology has been employed to capture accurate real-time transportation information to improve urban transportation planning. Recently, several studies introduced machine learning and deep learning techniques to investigate transportation utilization from multimodal sensors, including accelerometer, gyroscope, and magnetometer sensors. However, prior work has been constrained by impractical mobile computing with a large number of model parameters. We tackle this issue in this study by providing a hybrid deep learning model for identifying vehicle usages utilizing data from smartphone sensors. We conducted experiments on a publicly available dataset of human activity vibrations called the HAV dataset. The proposed model is evaluated with a variety of conventional deep learning algorithms. The performance assessment demonstrates that the proposed hybrid deep learning model classifies people's transportation behaviors more accurately than previous studies.
传感器先进技术促进了通过可穿戴设备识别人体运动的各种解决方案的发展。交通工具的特征已经成为智能交通系统中的有益应用,因为它可以为驾驶员辅助和智能交通管理等系统的实施提供上下文感知支持。智能手机传感技术被用于获取准确的实时交通信息,以改善城市交通规划。最近,一些研究引入了机器学习和深度学习技术来研究多模态传感器的交通利用率,包括加速度计、陀螺仪和磁力计传感器。然而,先前的工作受到不切实际的具有大量模型参数的移动计算的限制。在本研究中,我们通过提供一种混合深度学习模型来解决这个问题,该模型利用智能手机传感器的数据来识别车辆的使用情况。我们在一个公开的人类活动振动数据集上进行了实验,这个数据集叫做HAV数据集。用各种传统的深度学习算法对所提出的模型进行了评估。性能评估表明,所提出的混合深度学习模型比以往的研究更准确地分类了人们的交通行为。
{"title":"A Hybrid Deep Neural Network for Classifying Transportation Modes based on Human Activity Vibration","authors":"S. Mekruksavanich, Ponnipa Jantawong, I. You, A. Jitpattanakul","doi":"10.1109/KST53302.2022.9729079","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9729079","url":null,"abstract":"Sensor advanced technologies have facilitated the growth of various solutions for recognizing human movement through wearable devices. Characterization of the means of transportation has become beneficial applications in an intelligent transportation system since it enables context-aware support for the implementation of systems such as driver assistance and intelligent transportation management. Smartphone sensing technology has been employed to capture accurate real-time transportation information to improve urban transportation planning. Recently, several studies introduced machine learning and deep learning techniques to investigate transportation utilization from multimodal sensors, including accelerometer, gyroscope, and magnetometer sensors. However, prior work has been constrained by impractical mobile computing with a large number of model parameters. We tackle this issue in this study by providing a hybrid deep learning model for identifying vehicle usages utilizing data from smartphone sensors. We conducted experiments on a publicly available dataset of human activity vibrations called the HAV dataset. The proposed model is evaluated with a variety of conventional deep learning algorithms. The performance assessment demonstrates that the proposed hybrid deep learning model classifies people's transportation behaviors more accurately than previous studies.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123911852","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Loan Default Risk Prediction Using Knowledge Graph 基于知识图谱的贷款违约风险预测
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9729073
Md. Nurul Alam, M. Ali
Credit risk, also known as loan default risk, is one of the significant financial challenges in banking and financial institutions since it involves the uncertainty of the borrowers' ability to perform their contractual obligation. Banks and financial institutions rely on statistical and machine learning methods in predicting loan default to reduce the potential losses of issued loans. These machine learning applications may never achieve their full potential without the semantic context in the data. A knowledge graph is a collection of linked entities and objects that include semantic information to contextualize them. Knowledge graphs allow machines to incorporate human expertise into their decision-making and provide context to machine learning applications. Therefore, we proposed a loan default prediction model based on knowledge graph technology to improve the prediction model's accuracy and interpretability. The experimental results demonstrated that incorporating knowledge graph embedding as features can boost the performance of the conventional machine learning classifiers in predicting loan default risk.
信用风险,也被称为贷款违约风险,是银行和金融机构面临的重大金融挑战之一,因为它涉及到借款人履行合同义务能力的不确定性。银行和金融机构依靠统计和机器学习方法来预测贷款违约,以减少已发行贷款的潜在损失。如果没有数据中的语义上下文,这些机器学习应用程序可能永远无法充分发挥其潜力。知识图谱是一组相互关联的实体和对象的集合,这些实体和对象包含将它们上下文化的语义信息。知识图谱允许机器将人类的专业知识纳入其决策中,并为机器学习应用程序提供上下文。为此,我们提出了一种基于知识图技术的贷款违约预测模型,以提高预测模型的准确性和可解释性。实验结果表明,将知识图嵌入作为特征可以提高传统机器学习分类器在预测贷款违约风险方面的性能。
{"title":"Loan Default Risk Prediction Using Knowledge Graph","authors":"Md. Nurul Alam, M. Ali","doi":"10.1109/KST53302.2022.9729073","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9729073","url":null,"abstract":"Credit risk, also known as loan default risk, is one of the significant financial challenges in banking and financial institutions since it involves the uncertainty of the borrowers' ability to perform their contractual obligation. Banks and financial institutions rely on statistical and machine learning methods in predicting loan default to reduce the potential losses of issued loans. These machine learning applications may never achieve their full potential without the semantic context in the data. A knowledge graph is a collection of linked entities and objects that include semantic information to contextualize them. Knowledge graphs allow machines to incorporate human expertise into their decision-making and provide context to machine learning applications. Therefore, we proposed a loan default prediction model based on knowledge graph technology to improve the prediction model's accuracy and interpretability. The experimental results demonstrated that incorporating knowledge graph embedding as features can boost the performance of the conventional machine learning classifiers in predicting loan default risk.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"256 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121236876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Exploring Machine Learning Pipelines for Raman Spectral Classification of COVID-19 Samples 探索COVID-19样本拉曼光谱分类的机器学习管道
Pub Date : 2022-01-26 DOI: 10.1109/KST53302.2022.9729081
S. Deepaisarn, Chanvichet Vong, M. Perera
Raman Spectroscopy can analyze and identify the chemical compositions of samples. This study aims to develop a computational method based on machine learning algorithms to classify Raman spectra of serum samples from COVID-19 infected and non-infected human subjects. The method can potentially serve as a tool for rapid and accurate classification of COVID-19 versus non-COVID-19 patients and toward a direction for biomarker discoveries in research. Different machine learning classifiers were compared using pipelines with different dimensionality reduction and scaler techniques. The performance of each pipeline was investigated by varying the associate parameters. Assessment of dimensionality reduction application suggests that the pipelines generally performed better when the number of components does not exceed 50. The LightGBM model with ICA and MMScaler applied, yielded the highest test accuracy of 98.38% for pipelines with dimensionality reduction while the SVM model with MMScaler applied yielded the highest test accuracy of 96.77% for pipelines without dimensionality reduction. This study shows the effectiveness of Raman spectroscopy to classify COVID-19-induced characteristics in serum samples.
拉曼光谱可以分析和鉴定样品的化学成分。本研究旨在开发一种基于机器学习算法的计算方法,对COVID-19感染和未感染的人类血清样本进行拉曼光谱分类。该方法可以作为快速准确分类COVID-19与非COVID-19患者的工具,并为研究中发现生物标志物指明方向。不同的机器学习分类器使用管道与不同的降维和缩放技术进行比较。通过改变相关参数来研究每个管道的性能。对降维应用的评估表明,当组件数量不超过50时,管道的性能一般较好。应用ICA和MMScaler的LightGBM模型对降维管道的测试准确率最高,为98.38%;应用MMScaler的SVM模型对未降维管道的测试准确率最高,为96.77%。本研究证明了拉曼光谱对血清样品中covid -19诱导特征分类的有效性。
{"title":"Exploring Machine Learning Pipelines for Raman Spectral Classification of COVID-19 Samples","authors":"S. Deepaisarn, Chanvichet Vong, M. Perera","doi":"10.1109/KST53302.2022.9729081","DOIUrl":"https://doi.org/10.1109/KST53302.2022.9729081","url":null,"abstract":"Raman Spectroscopy can analyze and identify the chemical compositions of samples. This study aims to develop a computational method based on machine learning algorithms to classify Raman spectra of serum samples from COVID-19 infected and non-infected human subjects. The method can potentially serve as a tool for rapid and accurate classification of COVID-19 versus non-COVID-19 patients and toward a direction for biomarker discoveries in research. Different machine learning classifiers were compared using pipelines with different dimensionality reduction and scaler techniques. The performance of each pipeline was investigated by varying the associate parameters. Assessment of dimensionality reduction application suggests that the pipelines generally performed better when the number of components does not exceed 50. The LightGBM model with ICA and MMScaler applied, yielded the highest test accuracy of 98.38% for pipelines with dimensionality reduction while the SVM model with MMScaler applied yielded the highest test accuracy of 96.77% for pipelines without dimensionality reduction. This study shows the effectiveness of Raman spectroscopy to classify COVID-19-induced characteristics in serum samples.","PeriodicalId":433638,"journal":{"name":"2022 14th International Conference on Knowledge and Smart Technology (KST)","volume":"224 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134136967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2022 14th International Conference on Knowledge and Smart Technology (KST)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1