首页 > 最新文献

2021 RIVF International Conference on Computing and Communication Technologies (RIVF)最新文献

英文 中文
Fast pornographic video detection using Deep Learning 快速色情视频检测使用深度学习
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642154
Vinh-Nam Huynh, H. H. Nguyen
The recent rapid development of internet technology and applications leads to the booming of videos uploaded to and shared on the internet. However, some of them may contain impermissible content, especially pornographic videos, for the viewers. This problem raises a vast challenge in video filtering for many input videos. Concerning this matter, we introduce our system in which a key frame extraction method will be applied to the input video at the very first step. Subsequently, the Tensorflow object detection API is in charge of detecting and cropping any existing person in these frames. A Convolutional Neural Network (CNN) model then takes the cropped images and classifies them as pornography or not. The video is finally is marked valid for publishing according if the number of adult frames is below a threshold. Our experiments show that the proposed system can process videos much faster than human do while the accuracy is around 90% which can be meaningful to assist people in the task of video filtering.
近年来,互联网技术和应用的快速发展导致了视频上传和分享到互联网上的热潮。然而,其中一些可能包含不允许的内容,特别是色情视频,观众。这个问题对大量输入视频的视频过滤提出了巨大的挑战。针对这个问题,我们介绍了我们的系统,在系统的第一步,我们将对输入视频应用关键帧提取方法。随后,Tensorflow对象检测API负责检测和裁剪这些帧中的任何现有人物。然后,卷积神经网络(CNN)模型将裁剪后的图像分类为色情或非色情。如果视频的成人帧数低于阈值,则该视频最终被标记为有效发布。实验表明,该系统处理视频的速度比人类快得多,准确率在90%左右,对辅助人们进行视频过滤具有重要意义。
{"title":"Fast pornographic video detection using Deep Learning","authors":"Vinh-Nam Huynh, H. H. Nguyen","doi":"10.1109/RIVF51545.2021.9642154","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642154","url":null,"abstract":"The recent rapid development of internet technology and applications leads to the booming of videos uploaded to and shared on the internet. However, some of them may contain impermissible content, especially pornographic videos, for the viewers. This problem raises a vast challenge in video filtering for many input videos. Concerning this matter, we introduce our system in which a key frame extraction method will be applied to the input video at the very first step. Subsequently, the Tensorflow object detection API is in charge of detecting and cropping any existing person in these frames. A Convolutional Neural Network (CNN) model then takes the cropped images and classifies them as pornography or not. The video is finally is marked valid for publishing according if the number of adult frames is below a threshold. Our experiments show that the proposed system can process videos much faster than human do while the accuracy is around 90% which can be meaningful to assist people in the task of video filtering.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"12 23","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91405544","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Strengthening IDS against Evasion Attacks with GAN-based Adversarial Samples in SDN-enabled network sdn网络中基于gan的对抗样本增强IDS抗逃避攻击
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642111
Cao Phan Xuan Qui, Dang Hong Quang, Phan The Duy, Do Thi Thu Hien, V. Pham
With the spread of the number of smart devices in the context of Smart City, Software Defined Networking (SDN) is considered as a vital principle to manage a large-scale heterogeneous network within centralized controller. To deal with cyberattacks against such networks, intrusion detection system (IDS) is built to recognize and alert to the system administrator for further appropriate response. Currently, machine learning-based IDS (ML-IDS) has been explored and is still being developed. However, these systems give a high rate of false alert and are easily deceived by sophisticated attacks such as variants of attacks containing perturbation. Therefore, it is necessary to continuously evaluate and improve these systems by simulating mutation of real-world network attack. Relied on the Generative Discriminative Networks (GANs), we introduce DIGFuPAS, a framework that generates data flow of cyberattacks capable of bypassing ML-IDS. It can generate malicious data streams that mutate from real attack traffic making the IDS undetectable. The generated traffic flow is used to retrain ML-IDS, for improving the robustness of IDS in detecting sophisticated attacks. The experiments are performed and evaluated through 2 criteria: Detection rate (DR) and F1 Score (F1) on the public dataset, named CICIDS2017. DIGFuPAS can be used for continuously pentesting and evaluating IDS’s capability once integrated as an automated sustainability test pipeline for SDN-enabled networks.
随着智慧城市背景下智能设备数量的增加,软件定义网络(SDN)被认为是在集中控制器内管理大规模异构网络的重要原则。为了应对针对此类网络的网络攻击,我们建立了入侵检测系统(IDS),识别并提醒系统管理员采取进一步适当的应对措施。目前,基于机器学习的入侵检测(ML-IDS)已经被探索并仍在发展中。然而,这些系统的误报率很高,很容易被复杂的攻击所欺骗,比如包含扰动的攻击变体。因此,有必要通过模拟真实网络攻击的突变,对这些系统进行持续的评估和改进。基于生成判别网络(GANs),我们引入了DIGFuPAS,这是一个生成能够绕过ML-IDS的网络攻击数据流的框架。它可以从真实的攻击流量中生成恶意数据流,使IDS无法检测到。生成的流量流用于重新训练ML-IDS,以提高IDS检测复杂攻击的鲁棒性。在名为CICIDS2017的公共数据集上,通过检测率(Detection rate, DR)和F1评分(F1) 2个标准进行实验和评估。一旦集成为支持sdn的网络的自动化可持续性测试管道,DIGFuPAS可用于持续渗透测试和评估IDS的能力。
{"title":"Strengthening IDS against Evasion Attacks with GAN-based Adversarial Samples in SDN-enabled network","authors":"Cao Phan Xuan Qui, Dang Hong Quang, Phan The Duy, Do Thi Thu Hien, V. Pham","doi":"10.1109/RIVF51545.2021.9642111","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642111","url":null,"abstract":"With the spread of the number of smart devices in the context of Smart City, Software Defined Networking (SDN) is considered as a vital principle to manage a large-scale heterogeneous network within centralized controller. To deal with cyberattacks against such networks, intrusion detection system (IDS) is built to recognize and alert to the system administrator for further appropriate response. Currently, machine learning-based IDS (ML-IDS) has been explored and is still being developed. However, these systems give a high rate of false alert and are easily deceived by sophisticated attacks such as variants of attacks containing perturbation. Therefore, it is necessary to continuously evaluate and improve these systems by simulating mutation of real-world network attack. Relied on the Generative Discriminative Networks (GANs), we introduce DIGFuPAS, a framework that generates data flow of cyberattacks capable of bypassing ML-IDS. It can generate malicious data streams that mutate from real attack traffic making the IDS undetectable. The generated traffic flow is used to retrain ML-IDS, for improving the robustness of IDS in detecting sophisticated attacks. The experiments are performed and evaluated through 2 criteria: Detection rate (DR) and F1 Score (F1) on the public dataset, named CICIDS2017. DIGFuPAS can be used for continuously pentesting and evaluating IDS’s capability once integrated as an automated sustainability test pipeline for SDN-enabled networks.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"64 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83949433","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Interactive Z-line segmentation tool for Upper Gastrointestinal Endoscopy Images using Binary Partition Tree and U-net 基于二叉分割树和U-net的上消化道内镜图像交互式z线分割工具
Pub Date : 2021-08-19 DOI: 10.1109/RIVF51545.2021.9642141
X. Manh, Hai Vu, Xuan Dung Nguyen, Linh Hoang Pham Tu, V. Dao, Phuc Binh Nguyen, M. Nguyen
Z-line is a junction between esophageal and gastric mucosa which is an important landmark in exploring esophageal diseases such as Gastroesophageal Reflux Diseases (GERD). This paper describes an effective interactive segmentation tool for Z-line annotation from Upper Gastrointestinal Endoscopy (UGIE) images. To this end, we propose a method containing of two main steps: firstly, a coarse scheme is designed to roughly segment boundary regions of Z-line. Thanks to recent advances of deep neural networks in biomedical imaging such as U-net segmentation, Z-line annotation is automatically achieved with acceptable results. However, the U-net’s segmentation is not accurate enough due to gastric mucosa complexity. We then propose a fine-tuning scheme, which aims to prune the U-net’s results. The proposed method is based on Binary Partition Tree (BPT) algorithms, which BPT is built-in into a Graphic User Interface. Objective of the proposed framework is to help endoscopy doctors achieve the best segmentation results with lowest efforts of interactions via the GUI. The experiment was setup to evaluate effectiveness of the proposed method by comparing performances of four different segmentation schemes. They are manual segmentation by hand, fully automation by U-net, the interactive segmentation via BPT only, and the proposed scheme (U-net+BPT). The results confirmed that the proposed method converged faster to ideal regions than the other three. It took the lowest time costs and users’ efforts but achieved the best accuracy. The proposed method also suggest a feasible solution for segmenting abnormal regions in UGIE images.
z线是食管与胃粘膜的连接点,是探索胃食管反流病等食道疾病的重要标志。本文描述了一种有效的交互式分割工具,用于上消化道内镜(UGIE)图像的z线标注。为此,我们提出了一种包含两个主要步骤的方法:首先,设计一个粗略分割z线边界区域的粗格式;由于深度神经网络在生物医学成像中的最新进展,如U-net分割,z线标注可以自动实现,并且结果可以接受。然而,由于胃粘膜的复杂性,U-net的分割不够准确。然后,我们提出了一个微调方案,旨在减少U-net的结果。该方法基于二叉分割树(Binary Partition Tree, BPT)算法,该算法内置于图形用户界面中。提出的框架的目的是帮助内窥镜医生通过GUI以最少的交互努力获得最佳分割结果。通过比较四种不同分割方案的性能,建立了实验来评估该方法的有效性。它们分别是手工分割、U-net完全自动化分割、仅使用BPT的交互式分割以及提出的U-net+BPT方案。结果表明,该方法收敛到理想区域的速度比其他三种方法快。它花费了最少的时间成本和用户的努力,却达到了最好的精度。该方法也为UGIE图像中异常区域的分割提供了一种可行的解决方案。
{"title":"Interactive Z-line segmentation tool for Upper Gastrointestinal Endoscopy Images using Binary Partition Tree and U-net","authors":"X. Manh, Hai Vu, Xuan Dung Nguyen, Linh Hoang Pham Tu, V. Dao, Phuc Binh Nguyen, M. Nguyen","doi":"10.1109/RIVF51545.2021.9642141","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642141","url":null,"abstract":"Z-line is a junction between esophageal and gastric mucosa which is an important landmark in exploring esophageal diseases such as Gastroesophageal Reflux Diseases (GERD). This paper describes an effective interactive segmentation tool for Z-line annotation from Upper Gastrointestinal Endoscopy (UGIE) images. To this end, we propose a method containing of two main steps: firstly, a coarse scheme is designed to roughly segment boundary regions of Z-line. Thanks to recent advances of deep neural networks in biomedical imaging such as U-net segmentation, Z-line annotation is automatically achieved with acceptable results. However, the U-net’s segmentation is not accurate enough due to gastric mucosa complexity. We then propose a fine-tuning scheme, which aims to prune the U-net’s results. The proposed method is based on Binary Partition Tree (BPT) algorithms, which BPT is built-in into a Graphic User Interface. Objective of the proposed framework is to help endoscopy doctors achieve the best segmentation results with lowest efforts of interactions via the GUI. The experiment was setup to evaluate effectiveness of the proposed method by comparing performances of four different segmentation schemes. They are manual segmentation by hand, fully automation by U-net, the interactive segmentation via BPT only, and the proposed scheme (U-net+BPT). The results confirmed that the proposed method converged faster to ideal regions than the other three. It took the lowest time costs and users’ efforts but achieved the best accuracy. The proposed method also suggest a feasible solution for segmenting abnormal regions in UGIE images.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"49 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74364310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
VinaFood21: A Novel Dataset for Evaluating Vietnamese Food Recognition VinaFood21:一个评估越南食物识别的新数据集
Pub Date : 2021-08-06 DOI: 10.1109/RIVF51545.2021.9642151
Trong-Thuan Nguyen, Thuan Q. Nguyen, D. Vo, Vien Nguyen, Ngoc Ho, Nguyen D. Vo, Kiet Van Nguyen, Khang Nguyen
Vietnam is such an attractive tourist destination with its stunning and pristine landscapes and its top-rated unique food and drink. Among thousands of Vietnamese dishes, foreigners and native people are interested in easy-to-eat tastes and easy-to-do recipes, along with reasonable prices, mouthwatering flavors, and popularity. Due to the diversity and almost all the dishes have significant similarities and the lack of quality Vietnamese food datasets, it is hard to implement an auto system to classify Vietnamese food, therefore, make people easier to discover Vietnamese food. This paper introduces a new Vietnamese food dataset named VinaFood21, which consists of 13,950 images corresponding to 21 dishes. We use 10,044 images for model training and 6,682 test images to classify each food in the VinaFood21 dataset and achieved an average accuracy of 74.81% when fine-tuning CNN EfficientNet-B0.
越南是一个极具吸引力的旅游目的地,拥有令人惊叹的原始景观和一流的独特食物和饮料。在成千上万的越南菜中,外国人和当地人感兴趣的是容易吃的味道和容易做的食谱,以及合理的价格、令人垂涎的味道和受欢迎的程度。由于越南食物种类繁多,几乎所有的菜肴都有明显的相似之处,而且缺乏高质量的越南食物数据集,因此很难实现对越南食物进行自动分类的系统,因此,让人们更容易发现越南食物。本文介绍了一个新的越南美食数据集VinaFood21,该数据集由21道菜对应的13950张图片组成。我们使用10044张图像进行模型训练,6682张测试图像对VinaFood21数据集中的每种食物进行分类,在对CNN EfficientNet-B0进行微调后,平均准确率达到74.81%。
{"title":"VinaFood21: A Novel Dataset for Evaluating Vietnamese Food Recognition","authors":"Trong-Thuan Nguyen, Thuan Q. Nguyen, D. Vo, Vien Nguyen, Ngoc Ho, Nguyen D. Vo, Kiet Van Nguyen, Khang Nguyen","doi":"10.1109/RIVF51545.2021.9642151","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642151","url":null,"abstract":"Vietnam is such an attractive tourist destination with its stunning and pristine landscapes and its top-rated unique food and drink. Among thousands of Vietnamese dishes, foreigners and native people are interested in easy-to-eat tastes and easy-to-do recipes, along with reasonable prices, mouthwatering flavors, and popularity. Due to the diversity and almost all the dishes have significant similarities and the lack of quality Vietnamese food datasets, it is hard to implement an auto system to classify Vietnamese food, therefore, make people easier to discover Vietnamese food. This paper introduces a new Vietnamese food dataset named VinaFood21, which consists of 13,950 images corresponding to 21 dishes. We use 10,044 images for model training and 6,682 test images to classify each food in the VinaFood21 dataset and achieved an average accuracy of 74.81% when fine-tuning CNN EfficientNet-B0.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"12 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-08-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74723276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Automatically Detecting Cyberbullying Comments on Online Game Forums 自动检测网络游戏论坛上的网络欺凌评论
Pub Date : 2021-06-03 DOI: 10.1109/RIVF51545.2021.9642116
Hanh Hong-Phuc Vo, H. Tran, Son T. Luu
Online game forums are popular to most of game players. They use it to communicate and discuss the strategy of the game, or even to make friends. However, game forums also contain abusive and harassment speech, disturbing and threatening players. Therefore, it is necessary to automatically detect and remove cyberbullying comments to keep the game forum clean and friendly. We use the Cyberbullying dataset collected from World of Warcraft (WoW) and League of Legends (LoL) forums and train classification models to automatically detect whether a comment of a player is abusive or not. The result obtains 82.69% of macro F1-score for LoL forum and 83.86% of macro F1-score for WoW forum by the Toxic-BERT model on the Cyberbullying dataset.
网络游戏论坛深受广大游戏玩家的欢迎。他们用它来交流和讨论游戏策略,甚至是交朋友。然而,游戏论坛也包含辱骂和骚扰言论,扰乱和威胁玩家。因此,有必要自动检测和删除网络欺凌评论,以保持游戏论坛的清洁和友好。我们使用从魔兽世界(WoW)和英雄联盟(LoL)论坛收集的网络欺凌数据集和训练分类模型来自动检测玩家的评论是否辱骂。结果在网络欺凌数据集上使用Toxic-BERT模型得到LoL论坛的宏观f1得分为82.69%,WoW论坛的宏观f1得分为83.86%。
{"title":"Automatically Detecting Cyberbullying Comments on Online Game Forums","authors":"Hanh Hong-Phuc Vo, H. Tran, Son T. Luu","doi":"10.1109/RIVF51545.2021.9642116","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642116","url":null,"abstract":"Online game forums are popular to most of game players. They use it to communicate and discuss the strategy of the game, or even to make friends. However, game forums also contain abusive and harassment speech, disturbing and threatening players. Therefore, it is necessary to automatically detect and remove cyberbullying comments to keep the game forum clean and friendly. We use the Cyberbullying dataset collected from World of Warcraft (WoW) and League of Legends (LoL) forums and train classification models to automatically detect whether a comment of a player is abusive or not. The result obtains 82.69% of macro F1-score for LoL forum and 83.86% of macro F1-score for WoW forum by the Toxic-BERT model on the Cyberbullying dataset.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"52 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2021-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82014066","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Multimodal Fusion with BERT and Attention Mechanism for Fake News Detection 基于BERT的多模态融合及注意机制的假新闻检测
Pub Date : 2021-04-23 DOI: 10.1109/RIVF51545.2021.9642125
Nguyen Manh Duc Tuan, Pham Quang Nhat Minh
Fake news detection is an important task for in- creasing the reliability of the information on the internet since fake news is spreading fast on social media and has a negative effect on our society. In this paper, we present a novel method for detecting fake news by fusing multi-modal features derived from textual and visual data. Specifically, we proposed a scaled dot- product attention mechanism to capture the relationship between text features extracted by a pre-trained BERT model and visual features extracted by a pre-trained VGG-19 model. Experimental results showed that our method improved against the current state-of-the-art method on a public Twitter dataset by 3.1% accuracy.
假新闻检测是提高互联网信息可靠性的一项重要任务,因为假新闻在社交媒体上迅速传播,对我们的社会产生了负面影响。在本文中,我们提出了一种通过融合文本和视觉数据衍生的多模态特征来检测假新闻的新方法。具体而言,我们提出了一种缩放的点积注意机制来捕获预训练BERT模型提取的文本特征与预训练VGG-19模型提取的视觉特征之间的关系。实验结果表明,我们的方法在公共Twitter数据集上的准确率比目前最先进的方法提高了3.1%。
{"title":"Multimodal Fusion with BERT and Attention Mechanism for Fake News Detection","authors":"Nguyen Manh Duc Tuan, Pham Quang Nhat Minh","doi":"10.1109/RIVF51545.2021.9642125","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642125","url":null,"abstract":"Fake news detection is an important task for in- creasing the reliability of the information on the internet since fake news is spreading fast on social media and has a negative effect on our society. In this paper, we present a novel method for detecting fake news by fusing multi-modal features derived from textual and visual data. Specifically, we proposed a scaled dot- product attention mechanism to capture the relationship between text features extracted by a pre-trained BERT model and visual features extracted by a pre-trained VGG-19 model. Experimental results showed that our method improved against the current state-of-the-art method on a public Twitter dataset by 3.1% accuracy.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"24 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2021-04-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76306462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
An Empirical Study for Vietnamese Constituency Parsing with Pre-training 基于预训练的越南语选区分析实证研究
Pub Date : 2020-10-19 DOI: 10.1109/RIVF51545.2021.9642143
Tuan-Vi Tran, Xuan-Thien Pham, Duc-Vu Nguyen, Kiet Van Nguyen, N. Nguyen
Constituency parsing is an important task that gets more attention in natural language processing. In this work, we use a span-based approach for Vietnamese constituency parsing. Our method follows the self-attention encoder architecture and a chart decoder using a CKY-style inference algorithm. We present analyses of the experiment results of the comparison of our empirical method using pre-training models XLM-R and PhoBERT on both Vietnamese datasets VietTreebank and NIIVTB1. The results show that our model with XLM-R archived the significantly F1-score better than other pre-training models, VietTreebank at 81.19% and NIIVTB1 at 85.70%.
成分分析是自然语言处理中备受关注的一项重要任务。在这项工作中,我们使用基于跨度的方法进行越南语选区解析。我们的方法遵循自关注编码器架构和使用cky风格推理算法的图表解码器。我们在越南数据集VietTreebank和NIIVTB1上使用预训练模型XLM-R和PhoBERT对我们的经验方法的实验结果进行了比较分析。结果表明,我们的XLM-R模型的f1得分明显优于其他预训练模型,VietTreebank为81.19%,NIIVTB1为85.70%。
{"title":"An Empirical Study for Vietnamese Constituency Parsing with Pre-training","authors":"Tuan-Vi Tran, Xuan-Thien Pham, Duc-Vu Nguyen, Kiet Van Nguyen, N. Nguyen","doi":"10.1109/RIVF51545.2021.9642143","DOIUrl":"https://doi.org/10.1109/RIVF51545.2021.9642143","url":null,"abstract":"Constituency parsing is an important task that gets more attention in natural language processing. In this work, we use a span-based approach for Vietnamese constituency parsing. Our method follows the self-attention encoder architecture and a chart decoder using a CKY-style inference algorithm. We present analyses of the experiment results of the comparison of our empirical method using pre-training models XLM-R and PhoBERT on both Vietnamese datasets VietTreebank and NIIVTB1. The results show that our model with XLM-R archived the significantly F1-score better than other pre-training models, VietTreebank at 81.19% and NIIVTB1 at 85.70%.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"52 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2020-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84687850","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Diagnosis of heart disease patients using fuzzy classification technique 模糊分类技术在心脏病诊断中的应用
Pub Date : 2014-12-01 DOI: 10.1109/ICCCT2.2014.7066746
V. Krishnaiah, M. Srinivas, G. Narsimha, N. S. Chandra
Data mining technique in the history of medical data found with enormous investigations found that the prediction of heart disease is very important in medical science. In medical history it is observed that the unstructured data as heterogeneous data and it is observed that the data formed with different attributes should be analyzed to predict and provide information for making diagnosis of a heart patient. Various techniques in Data Mining have been applied to predict the heart disease patients. But, the uncertainty in data was not removed with the techniques available in data mining and implemented by various authors. To remove uncertainty of unstructured data, an attempt was made by introducing fuzziness in the measured data. A membership function was designed and incorporated with the measured value to remove uncertainty and fuzzified data was used to predict the heart disease patients.. Further, an attempt was made to classify the patients based on the attributes collected from medical field. Minimum Euclidean distance Fuzzy K-NN classifier was designed to classify the training and testing data belonging to different classes. It was found that Fuzzy K-NN classifier suits well as compared with other classifiers of parametric techniques.
数据挖掘技术在医学数据发现的历史上,通过大量的调查发现,心脏疾病的预测在医学科学中非常重要。在病史中,将非结构化数据视为异构数据,对具有不同属性的数据进行分析,预测并为心脏病患者的诊断提供信息。数据挖掘中的各种技术已被应用于心脏病患者的预测。但是,数据中的不确定性并没有被数据挖掘中可用的技术和各种作者实现的技术所消除。为了消除非结构化数据的不确定性,尝试在测量数据中引入模糊性。设计了隶属函数,并与测量值相结合,消除了不确定性,利用模糊数据对心脏病患者进行预测。在此基础上,尝试基于医学领域收集的属性对患者进行分类。设计了最小欧氏距离模糊K-NN分类器,对不同类别的训练数据和测试数据进行分类。与其他参数化分类器相比,模糊K-NN分类器具有较好的适用性。
{"title":"Diagnosis of heart disease patients using fuzzy classification technique","authors":"V. Krishnaiah, M. Srinivas, G. Narsimha, N. S. Chandra","doi":"10.1109/ICCCT2.2014.7066746","DOIUrl":"https://doi.org/10.1109/ICCCT2.2014.7066746","url":null,"abstract":"Data mining technique in the history of medical data found with enormous investigations found that the prediction of heart disease is very important in medical science. In medical history it is observed that the unstructured data as heterogeneous data and it is observed that the data formed with different attributes should be analyzed to predict and provide information for making diagnosis of a heart patient. Various techniques in Data Mining have been applied to predict the heart disease patients. But, the uncertainty in data was not removed with the techniques available in data mining and implemented by various authors. To remove uncertainty of unstructured data, an attempt was made by introducing fuzziness in the measured data. A membership function was designed and incorporated with the measured value to remove uncertainty and fuzzified data was used to predict the heart disease patients.. Further, an attempt was made to classify the patients based on the attributes collected from medical field. Minimum Euclidean distance Fuzzy K-NN classifier was designed to classify the training and testing data belonging to different classes. It was found that Fuzzy K-NN classifier suits well as compared with other classifiers of parametric techniques.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"26 11 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77263630","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Scalable load balancing using virtualization based on approximation 使用基于近似的虚拟化的可扩展负载平衡
Pub Date : 2014-12-01 DOI: 10.1109/ICCCT2.2014.7066720
Mohammed A. Saifullah, M. A. Maluk Mohammed
The number of users and services on Internet are increasing day by day resulting in high traffic and load on the web servers. This is in turn increasing the service time of web requests and degrading the quality of service. A well known solution to this problem is replication of content using cluster of web servers. An efficient server load balancing policy is required to achieve scalability and high performance of the service offered by cluster of web servers. Under dynamic, secure and database driven loads, existing load balancing strategies are suffering from performance degradation. In this paper, we proposed Scalable Load Balancing using Virtualization based on Approximation Algorithm (SLBVA). SLBVA is an estimation strategy as it is challenging to correctly measure the load on each web server of a cluster. SLBVA algorithm is capable of offering guarantees for different client priorities, such as premium customers and default customers. We show that using SLBVA strategy web servers are able to maintain Service Level Agreements (SLA) without the need of a priori over-dimensioning of server resources. This is achieved by taking the real perspective of the service requests using the measurement of arrival rates at that time and judiciously discard some requests from the default clients if the default customers traffic is high. If the arrival rate of premium customers goes beyond the capacity of cluster, we will increase the capacity of cluster using virtualization by utilizing the unused servers from the under-utilized server farms. We analyzed and compared the experimental results of SLBVA algorithm with the results of the very popular load balancing algorithm, Weighted Round Robin (WRR). We show that even though the SLBVA strategy takes a little more server processing resources than WRR, it is capable to render assurances unlike WRR.
互联网上的用户和服务数量日益增加,导致网络服务器的流量和负载越来越大。这反过来又增加了web请求的服务时间,降低了服务质量。这个问题的一个众所周知的解决方案是使用web服务器集群复制内容。为了实现web服务器集群提供的服务的可伸缩性和高性能,需要一个有效的服务器负载平衡策略。在动态、安全和数据库驱动的负载下,现有的负载平衡策略存在性能下降的问题。本文提出了基于近似算法(SLBVA)的虚拟化可扩展负载均衡。SLBVA是一种评估策略,因为正确测量集群的每个web服务器上的负载具有挑战性。SLBVA算法能够为不同的客户端优先级提供保证,例如高级客户和默认客户。我们表明,使用SLBVA策略的web服务器能够维护服务水平协议(SLA),而不需要先验的服务器资源多维化。这是通过使用当时的到达率度量来获取服务请求的真实视图来实现的,并且如果默认客户流量很高,则明智地丢弃来自默认客户端的一些请求。如果高级客户的到达率超过集群的容量,我们将通过利用未充分利用的服务器群中未使用的服务器来使用虚拟化来增加集群的容量。我们将SLBVA算法的实验结果与非常流行的负载均衡算法加权轮询(WRR)的实验结果进行了分析和比较。我们表明,尽管SLBVA策略比WRR需要更多的服务器处理资源,但它能够提供与WRR不同的保证。
{"title":"Scalable load balancing using virtualization based on approximation","authors":"Mohammed A. Saifullah, M. A. Maluk Mohammed","doi":"10.1109/ICCCT2.2014.7066720","DOIUrl":"https://doi.org/10.1109/ICCCT2.2014.7066720","url":null,"abstract":"The number of users and services on Internet are increasing day by day resulting in high traffic and load on the web servers. This is in turn increasing the service time of web requests and degrading the quality of service. A well known solution to this problem is replication of content using cluster of web servers. An efficient server load balancing policy is required to achieve scalability and high performance of the service offered by cluster of web servers. Under dynamic, secure and database driven loads, existing load balancing strategies are suffering from performance degradation. In this paper, we proposed Scalable Load Balancing using Virtualization based on Approximation Algorithm (SLBVA). SLBVA is an estimation strategy as it is challenging to correctly measure the load on each web server of a cluster. SLBVA algorithm is capable of offering guarantees for different client priorities, such as premium customers and default customers. We show that using SLBVA strategy web servers are able to maintain Service Level Agreements (SLA) without the need of a priori over-dimensioning of server resources. This is achieved by taking the real perspective of the service requests using the measurement of arrival rates at that time and judiciously discard some requests from the default clients if the default customers traffic is high. If the arrival rate of premium customers goes beyond the capacity of cluster, we will increase the capacity of cluster using virtualization by utilizing the unused servers from the under-utilized server farms. We analyzed and compared the experimental results of SLBVA algorithm with the results of the very popular load balancing algorithm, Weighted Round Robin (WRR). We show that even though the SLBVA strategy takes a little more server processing resources than WRR, it is capable to render assurances unlike WRR.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"53 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81268264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Byte Rotation Encryption Algorithm through parallel processing and multi-core utilization 字节旋转加密算法通过并行处理和多核利用
Pub Date : 2014-12-01 DOI: 10.1109/ICCCT2.2014.7066719
G. Thirumaleswari, Ch. Suneetha, G. C. Bharathi
Securing digital data has become tedious task as the technology is increasing. Existing encryption algorithms such as AES,DES and Blowfish ensure information security but consume lot of time as the security level increases. In this paper, Byte Rotation Encryption Algorithm (BREA) has implemented using parallel processing and multi-core utilization. BREA divides data into fixed size blocks. Each block is processed parallely using random number key. So the set of blocks are executed in parallel by utilizing all the available CPU cores. Finally, from the experimental analysis, it is observed that the proposed BREA algorithm reduces execution time when the number of cores has increased.
随着技术的发展,保护数字数据已成为一项繁琐的任务。现有的AES、DES、Blowfish等加密算法虽然能够保证信息的安全,但随着安全级别的提高,这些算法会消耗大量的时间。本文采用并行处理和多核利用的方法实现了字节旋转加密算法(BREA)。BREA将数据分成固定大小的块。每个区块使用随机数键并行处理。因此,通过利用所有可用的CPU内核并行执行这组块。最后,通过实验分析发现,随着核数的增加,BREA算法的执行时间有所减少。
{"title":"Byte Rotation Encryption Algorithm through parallel processing and multi-core utilization","authors":"G. Thirumaleswari, Ch. Suneetha, G. C. Bharathi","doi":"10.1109/ICCCT2.2014.7066719","DOIUrl":"https://doi.org/10.1109/ICCCT2.2014.7066719","url":null,"abstract":"Securing digital data has become tedious task as the technology is increasing. Existing encryption algorithms such as AES,DES and Blowfish ensure information security but consume lot of time as the security level increases. In this paper, Byte Rotation Encryption Algorithm (BREA) has implemented using parallel processing and multi-core utilization. BREA divides data into fixed size blocks. Each block is processed parallely using random number key. So the set of blocks are executed in parallel by utilizing all the available CPU cores. Finally, from the experimental analysis, it is observed that the proposed BREA algorithm reduces execution time when the number of cores has increased.","PeriodicalId":6860,"journal":{"name":"2021 RIVF International Conference on Computing and Communication Technologies (RIVF)","volume":"78 1","pages":"1-4"},"PeriodicalIF":0.0,"publicationDate":"2014-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76886974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2021 RIVF International Conference on Computing and Communication Technologies (RIVF)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1