首页 > 最新文献

2022 14th International Conference on Knowledge and Systems Engineering (KSE)最新文献

英文 中文
Detecting Coincidental Correctness and Mitigating Its Impacts on Localizing Variability Faults 巧合正确性检测及其对局部变异性故障的影响
Pub Date : 2022-10-19 DOI: 10.1109/KSE56063.2022.9953777
Thu-Trang Nguyen, H. Vo
Coincidental correctness is the phenomenon that test cases execute faulty statements yet still produce correct/expected outputs. In software testing, this problem is prevalent and causes negative impacts on fault localization performance. Although detecting coincidentally correct (CC) tests and mitigating their impacts on localizing faults in non-configurable systems have been studied in-depth, handling CC tests in Software Product Line (SPL) systems have been unexplored. To test an SPL system, products are often sampled, and each product is tested individually. The CC test cases, that occur in the test suite of a product, not only affect the testing results of the corresponding product but also affect the overall testing results of the system. This could negatively affect fault localization performance and decelerate the quality assurance process for the system. In this paper, we introduce DEMiC, a novel approach to detect CC tests and mitigate their impacts on localizing variability faults in SPL systems. Our key idea to detect CC tests is that two similar tests tend to examine similar behaviors of the system and should have a similar testing state (i.e., both passed or failed). If only one of them failed, the other could be coincidentally passed. In addition, we propose several solutions to mitigate the negative impacts of CC tests on variability fault localization at different levels. Our experimental results on +2,6M test cases of five widely used SPL systems show that DEMiC can effectively detect CC tests, with 97% accuracy on average. In addition, DEMiC could help to improve the fault localization performance by 61%.
巧合正确性是测试用例执行错误语句但仍然产生正确/预期输出的现象。在软件测试中,这个问题非常普遍,并且会对故障定位性能造成负面影响。尽管在非可配置系统中检测巧合正确(CC)测试并减轻其对故障定位的影响已经得到了深入的研究,但在软件产品线(SPL)系统中处理CC测试尚未得到探索。为了测试SPL系统,通常对产品进行取样,并且对每个产品进行单独测试。CC测试用例发生在产品的测试套件中,它不仅会影响相应产品的测试结果,还会影响系统的整体测试结果。这可能会对故障定位性能产生负面影响,并减慢系统的质量保证过程。在本文中,我们介绍了一种检测CC测试并减轻其对SPL系统局部变性故障影响的新方法。我们检测CC测试的关键思想是,两个相似的测试倾向于检查系统的相似行为,并且应该具有相似的测试状态(即,通过或失败)。如果其中只有一个不及格,另一个可能会碰巧通过。此外,针对CC测试在不同层次上对变异性故障定位的负面影响,提出了不同的解决方案。我们在5个广泛使用的SPL系统的+ 2,600万个测试用例上的实验结果表明,DEMiC可以有效地检测CC测试,平均准确率为97%。此外,DEMiC还可以将故障定位性能提高61%。
{"title":"Detecting Coincidental Correctness and Mitigating Its Impacts on Localizing Variability Faults","authors":"Thu-Trang Nguyen, H. Vo","doi":"10.1109/KSE56063.2022.9953777","DOIUrl":"https://doi.org/10.1109/KSE56063.2022.9953777","url":null,"abstract":"Coincidental correctness is the phenomenon that test cases execute faulty statements yet still produce correct/expected outputs. In software testing, this problem is prevalent and causes negative impacts on fault localization performance. Although detecting coincidentally correct (CC) tests and mitigating their impacts on localizing faults in non-configurable systems have been studied in-depth, handling CC tests in Software Product Line (SPL) systems have been unexplored. To test an SPL system, products are often sampled, and each product is tested individually. The CC test cases, that occur in the test suite of a product, not only affect the testing results of the corresponding product but also affect the overall testing results of the system. This could negatively affect fault localization performance and decelerate the quality assurance process for the system. In this paper, we introduce DEMiC, a novel approach to detect CC tests and mitigate their impacts on localizing variability faults in SPL systems. Our key idea to detect CC tests is that two similar tests tend to examine similar behaviors of the system and should have a similar testing state (i.e., both passed or failed). If only one of them failed, the other could be coincidentally passed. In addition, we propose several solutions to mitigate the negative impacts of CC tests on variability fault localization at different levels. Our experimental results on +2,6M test cases of five widely used SPL systems show that DEMiC can effectively detect CC tests, with 97% accuracy on average. In addition, DEMiC could help to improve the fault localization performance by 61%.","PeriodicalId":330865,"journal":{"name":"2022 14th International Conference on Knowledge and Systems Engineering (KSE)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131960456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distill Knowledge in Multi-task Reinforcement Learning with Optimal-Transport Regularization 基于最优传输正则化的多任务强化学习中的知识提取
Pub Date : 2022-10-19 DOI: 10.1109/KSE56063.2022.9953750
Bang Giang Le, Viet-Cuong Ta
In multi-task reinforcement learning, it is possible to improve the data efficiency of training agents by transferring knowledge from other different but related tasks. Because the experiences from different tasks are usually biased toward the specific task goals. Traditional methods rely on Kullback-Leibler regularization to stabilize the transfer of knowledge from one task to the others. In this work, we explore the direction of replacing the Kullback-Leibler divergence with a novel Optimal transport-based regularization. By using the Sinkhorn mapping, we can approximate the Optimal transport distance between the state distribution of tasks. The distance is then used as an amortized reward to regularize the amount of sharing information. We experiment our frameworks on several grid-based navigation multi-goal to validate the effectiveness of the approach. The results show that our added Optimal transport-based rewards are able to speed up the learning process of agents and outperforms several baselines on multi-task learning.
在多任务强化学习中,可以通过从其他不同但相关的任务中转移知识来提高训练代理的数据效率。因为来自不同任务的经验通常会偏向于特定的任务目标。传统的方法依靠Kullback-Leibler正则化来稳定知识从一个任务到另一个任务的转移。在这项工作中,我们探索了用一种新的基于最优传输的正则化取代Kullback-Leibler散度的方向。通过使用Sinkhorn映射,我们可以近似出任务状态分布之间的最优传输距离。然后使用距离作为平摊奖励来正则化共享信息的数量。我们在几个基于网格的多目标导航上进行了实验,以验证该方法的有效性。结果表明,我们添加的基于最优运输的奖励能够加快智能体的学习过程,并且在多任务学习上优于几个基线。
{"title":"Distill Knowledge in Multi-task Reinforcement Learning with Optimal-Transport Regularization","authors":"Bang Giang Le, Viet-Cuong Ta","doi":"10.1109/KSE56063.2022.9953750","DOIUrl":"https://doi.org/10.1109/KSE56063.2022.9953750","url":null,"abstract":"In multi-task reinforcement learning, it is possible to improve the data efficiency of training agents by transferring knowledge from other different but related tasks. Because the experiences from different tasks are usually biased toward the specific task goals. Traditional methods rely on Kullback-Leibler regularization to stabilize the transfer of knowledge from one task to the others. In this work, we explore the direction of replacing the Kullback-Leibler divergence with a novel Optimal transport-based regularization. By using the Sinkhorn mapping, we can approximate the Optimal transport distance between the state distribution of tasks. The distance is then used as an amortized reward to regularize the amount of sharing information. We experiment our frameworks on several grid-based navigation multi-goal to validate the effectiveness of the approach. The results show that our added Optimal transport-based rewards are able to speed up the learning process of agents and outperforms several baselines on multi-task learning.","PeriodicalId":330865,"journal":{"name":"2022 14th International Conference on Knowledge and Systems Engineering (KSE)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126351942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Knowledge-based Problem Solving and Reasoning methods 基于知识的问题解决和推理方法
Pub Date : 2022-10-19 DOI: 10.1109/KSE56063.2022.9953617
N. Do, H. Nguyen
Intelligent Problem Solver (IPS) is an intelligent system for solving practical problems in the determined domain by using human knowledge. Thus, designing of the knowledge base and the inference engine of IPS systems are important. This study proposed a general model for knowledge representation by using a kernel ontology combining other knowledge components, called Integ-Ontology Based on this model, the model of problems is presented. The reasoning method is also proposed. This method includes the inference processing and techniques of heuristic rules, sample problems and pattern to speed up the problem solving. The Integ-Ontology and its reasoning method are applied to design practical IPS in solid geometry and Direct Current (DC) Electrical Circuits.
智能问题解决器(IPS)是一种利用人类知识解决确定领域内实际问题的智能系统。因此,IPS系统的知识库和推理引擎的设计就显得尤为重要。本研究提出了一种通用的知识表示模型,即利用核心本体结合其他知识组件来表示知识,并在此基础上提出了问题模型。并提出了推理方法。该方法包括推理处理和启发式规则、样本问题和模式技术,以加快问题的求解速度。将集成本体及其推理方法应用于实体几何和直流电路的实用IPS设计。
{"title":"Knowledge-based Problem Solving and Reasoning methods","authors":"N. Do, H. Nguyen","doi":"10.1109/KSE56063.2022.9953617","DOIUrl":"https://doi.org/10.1109/KSE56063.2022.9953617","url":null,"abstract":"Intelligent Problem Solver (IPS) is an intelligent system for solving practical problems in the determined domain by using human knowledge. Thus, designing of the knowledge base and the inference engine of IPS systems are important. This study proposed a general model for knowledge representation by using a kernel ontology combining other knowledge components, called Integ-Ontology Based on this model, the model of problems is presented. The reasoning method is also proposed. This method includes the inference processing and techniques of heuristic rules, sample problems and pattern to speed up the problem solving. The Integ-Ontology and its reasoning method are applied to design practical IPS in solid geometry and Direct Current (DC) Electrical Circuits.","PeriodicalId":330865,"journal":{"name":"2022 14th International Conference on Knowledge and Systems Engineering (KSE)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-10-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115070635","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Online pseudo labeling for polyp segmentation with momentum networks 基于动量网络的息肉分割在线伪标记
Pub Date : 2022-09-29 DOI: 10.1109/KSE56063.2022.9953785
Toan Pham Van, Linh Doan Bao, Thanh-Tung Nguyen, Duc Trung Tran, Q. Nguyen, D. V. Sang
emantic segmentation is an essential task in developing medical image diagnosis systems. However, building an annotated medical dataset is expensive. Thus, semi-supervised methods are significant in this circumstance. In semi-supervised learning, the quality of labels plays a crucial role in modelperformance. In this work, we present a new pseudo labeling strategy that enhances the quality of pseudo labels used for training student networks. We follow the multi-stage semi-supervised training approach, which trains a teacher model on a labeled dataset and then uses the trained teacher to render pseudo labels for student training. By doing so, the pseudo labels will be updated and more precise as training progress. The key difference between previous and our methods is that we update the teacher model during the student training process. So the quality of pseudo labels is improved during the student training process. We also propose a simple but effective strategy to enhance the quality of pseudo labels using a momentum model - a slow copy version of the original model during training. By applying the momentum model combined with re-rendering pseudo labels during student training, we achieved an average of 84.1% Dice Score on five datasets (i.e., Kvarsir, CVC-ClinicDB, ETIS-LaribPolypDB, CVC-ColonDB, and CVC-300) with only 20% of the dataset used as labeled data. Our results surpass common practice by 3% and even approach fully-supervised results on some datasets. Oursource code and pre-trained models are available at https://github.com/sun-asterisk-research/online_learning_ssl
Emantic分割是医学图像诊断系统开发中的一项重要任务。然而,构建一个带注释的医学数据集是昂贵的。因此,在这种情况下,半监督方法是重要的。在半监督学习中,标签的质量对模型的性能起着至关重要的作用。在这项工作中,我们提出了一种新的伪标签策略,可以提高用于训练学生网络的伪标签的质量。我们遵循多阶段半监督训练方法,该方法在标记数据集上训练教师模型,然后使用训练有素的教师为学生训练提供伪标签。通过这样做,伪标签将随着训练的进展而更新和更精确。我们的方法与以前的方法的关键区别在于,我们在学生培养过程中更新了教师模型。在学员训练过程中,提高了伪标签的质量。我们还提出了一种简单但有效的策略来提高伪标签的质量,使用动量模型-在训练过程中原始模型的缓慢复制版本。通过在学生训练期间应用结合重新渲染伪标签的momentum模型,我们在5个数据集(即Kvarsir、CVC-ClinicDB、ETIS-LaribPolypDB、CVC-ColonDB和CVC-300)上实现了平均84.1%的Dice Score,其中只有20%的数据集被用作标记数据。我们的结果比通常的做法高出3%,甚至在一些数据集上接近完全监督的结果。源代码和预训练模型可在https://github.com/sun-asterisk-research/online_learning_ssl上获得
{"title":"Online pseudo labeling for polyp segmentation with momentum networks","authors":"Toan Pham Van, Linh Doan Bao, Thanh-Tung Nguyen, Duc Trung Tran, Q. Nguyen, D. V. Sang","doi":"10.1109/KSE56063.2022.9953785","DOIUrl":"https://doi.org/10.1109/KSE56063.2022.9953785","url":null,"abstract":"emantic segmentation is an essential task in developing medical image diagnosis systems. However, building an annotated medical dataset is expensive. Thus, semi-supervised methods are significant in this circumstance. In semi-supervised learning, the quality of labels plays a crucial role in modelperformance. In this work, we present a new pseudo labeling strategy that enhances the quality of pseudo labels used for training student networks. We follow the multi-stage semi-supervised training approach, which trains a teacher model on a labeled dataset and then uses the trained teacher to render pseudo labels for student training. By doing so, the pseudo labels will be updated and more precise as training progress. The key difference between previous and our methods is that we update the teacher model during the student training process. So the quality of pseudo labels is improved during the student training process. We also propose a simple but effective strategy to enhance the quality of pseudo labels using a momentum model - a slow copy version of the original model during training. By applying the momentum model combined with re-rendering pseudo labels during student training, we achieved an average of 84.1% Dice Score on five datasets (i.e., Kvarsir, CVC-ClinicDB, ETIS-LaribPolypDB, CVC-ColonDB, and CVC-300) with only 20% of the dataset used as labeled data. Our results surpass common practice by 3% and even approach fully-supervised results on some datasets. Oursource code and pre-trained models are available at https://github.com/sun-asterisk-research/online_learning_ssl","PeriodicalId":330865,"journal":{"name":"2022 14th International Conference on Knowledge and Systems Engineering (KSE)","volume":"26 10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126037224","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Using Multiple Code Representations to Prioritize Static Analysis Warnings 使用多种代码表示来确定静态分析警告的优先级
Pub Date : 2022-09-25 DOI: 10.1109/KSE56063.2022.9953786
Thanh Vu, H. Vo
In order to ensure the quality of software and prevent attacks from hackers on critical systems, static analysis tools are frequently utilized to detect vulnerabilities in the early development phase. However, these tools often report a large number of warnings with a high false-positive rate, which causes many difficulties for developers. In this paper, we introduce VULRG, a novel approach to address this problem. Specifically, VuLRG predicts and ranks the warnings based on their likelihoods to be true positives. To predict these likelihoods, VuLRG combines two deep learning models CNN and BiGRU to capture the context of each warning in terms of program syntax, control flow, and program dependence. Our experimental results on a real-world dataset of 6,620 warnings show that VuLRG’s Recall at Top-50% is 90%. This means that using VuLRG, 90% of the vulnerabilities can be found by examining only 50% of the warnings. Moreover, at Top-5%, VULRG can improve the state-of-the-art approach by +30% in both Precision and Recall.
为了保证软件的质量和防止黑客对关键系统的攻击,静态分析工具经常被用于在早期开发阶段检测漏洞。然而,这些工具经常报告大量的警告,并且假阳性率很高,这给开发人员带来了许多困难。在本文中,我们介绍了一种解决这一问题的新方法——VULRG。具体来说,VuLRG根据它们成为真阳性的可能性对警告进行预测和排序。为了预测这些可能性,VuLRG结合了CNN和BiGRU两个深度学习模型,从程序语法、控制流和程序依赖性方面捕获每个警告的上下文。我们在包含6620个警告的真实数据集上的实验结果表明,在Top-50%时,VuLRG的召回率为90%。这意味着使用VuLRG,只需检查50%的警告就可以发现90%的漏洞。此外,在Top-5%的情况下,VULRG可以将最先进的方法在精度和召回率方面提高30%。
{"title":"Using Multiple Code Representations to Prioritize Static Analysis Warnings","authors":"Thanh Vu, H. Vo","doi":"10.1109/KSE56063.2022.9953786","DOIUrl":"https://doi.org/10.1109/KSE56063.2022.9953786","url":null,"abstract":"In order to ensure the quality of software and prevent attacks from hackers on critical systems, static analysis tools are frequently utilized to detect vulnerabilities in the early development phase. However, these tools often report a large number of warnings with a high false-positive rate, which causes many difficulties for developers. In this paper, we introduce VULRG, a novel approach to address this problem. Specifically, VuLRG predicts and ranks the warnings based on their likelihoods to be true positives. To predict these likelihoods, VuLRG combines two deep learning models CNN and BiGRU to capture the context of each warning in terms of program syntax, control flow, and program dependence. Our experimental results on a real-world dataset of 6,620 warnings show that VuLRG’s Recall at Top-50% is 90%. This means that using VuLRG, 90% of the vulnerabilities can be found by examining only 50% of the warnings. Moreover, at Top-5%, VULRG can improve the state-of-the-art approach by +30% in both Precision and Recall.","PeriodicalId":330865,"journal":{"name":"2022 14th International Conference on Knowledge and Systems Engineering (KSE)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121685274","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Non-Standard Vietnamese Word Detection and Normalization for Text–to–Speech 文本到语音的非标准越南语单词检测与规范化
Pub Date : 2022-09-07 DOI: 10.1109/KSE56063.2022.9953791
Huu-Tien Dang, Thi-Hai-Yen Vuong, X. Phan
Converting written texts into their spoken forms is an essential problem in any text-to-speech (TTS) systems. However, building an effective text normalization solution for a real-world TTS system face two main challenges: (1) the semantic ambiguity of non-standard words (NSWs), e.g., numbers, dates, ranges, scores, abbreviations, and (2) transforming NSWs into pronounceable syllables, such as URL, email address, hashtag, and contact name. In this paper, we propose a new two-phase normalization approach to deal with these challenges. First, a model-based tagger is designed to detect NSWs. Then, depending on NSW types, a rule-based normalizer expands those NSWs into their final verbal forms. We conducted three empirical experiments for NSW detection using Conditional Random Fields (CRFs), BiLSTM-CNN-CRF, and BERT-BiGRU-CRF models on a manually annotated dataset including 5819 sentences extracted from Vietnamese news articles. In the second phase, we propose a forward lexicon-based maximum matching algorithm to split down the hashtag, email, URL, and contact name. The experimental results of the tagging phase show that the average F1 scores of the BiLSTM-CNN-CRF and CRF models are above 90.00%, reaching the highest F1 of 95.00% with the BERT-BiGRU-CRF model. Overall, our approach has low sentence error rates, at 8.15% with CRF and 7.11% with BiLSTM-CNNCRF taggers, and only 6.67% with BERT-BiGRU-CRF tagger.
将书面文本转换为口语形式是任何文本到语音(TTS)系统的基本问题。然而,为现实世界的TTS系统构建有效的文本规范化解决方案面临两个主要挑战:(1)非标准单词(nsw)的语义歧义,如数字、日期、范围、分数、缩写;(2)将非标准单词(nsw)转换为可发音的音节,如URL、电子邮件地址、hashtag和联系人姓名。在本文中,我们提出了一种新的两阶段规范化方法来应对这些挑战。首先,设计了一个基于模型的标注器来检测nsw。然后,根据NSW类型,基于规则的规范化器将这些NSW扩展为最终的口头形式。本文采用条件随机场(CRFs)、BiLSTM-CNN-CRF和BERT-BiGRU-CRF模型,在越南新闻文章中提取的5819个句子的人工标注数据集上进行了三个NSW检测的实证实验。在第二阶段,我们提出了一种基于正向词典的最大匹配算法来拆分标签、电子邮件、URL和联系人姓名。标记阶段的实验结果表明,BiLSTM-CNN-CRF和CRF模型的平均F1分数都在90.00%以上,其中BERT-BiGRU-CRF模型的F1分数最高,达到95.00%。总体而言,我们的方法具有较低的句子错误率,使用CRF和BiLSTM-CNNCRF标注器的错误率分别为8.15%和7.11%,使用BERT-BiGRU-CRF标注器的错误率只有6.67%。
{"title":"Non-Standard Vietnamese Word Detection and Normalization for Text–to–Speech","authors":"Huu-Tien Dang, Thi-Hai-Yen Vuong, X. Phan","doi":"10.1109/KSE56063.2022.9953791","DOIUrl":"https://doi.org/10.1109/KSE56063.2022.9953791","url":null,"abstract":"Converting written texts into their spoken forms is an essential problem in any text-to-speech (TTS) systems. However, building an effective text normalization solution for a real-world TTS system face two main challenges: (1) the semantic ambiguity of non-standard words (NSWs), e.g., numbers, dates, ranges, scores, abbreviations, and (2) transforming NSWs into pronounceable syllables, such as URL, email address, hashtag, and contact name. In this paper, we propose a new two-phase normalization approach to deal with these challenges. First, a model-based tagger is designed to detect NSWs. Then, depending on NSW types, a rule-based normalizer expands those NSWs into their final verbal forms. We conducted three empirical experiments for NSW detection using Conditional Random Fields (CRFs), BiLSTM-CNN-CRF, and BERT-BiGRU-CRF models on a manually annotated dataset including 5819 sentences extracted from Vietnamese news articles. In the second phase, we propose a forward lexicon-based maximum matching algorithm to split down the hashtag, email, URL, and contact name. The experimental results of the tagging phase show that the average F1 scores of the BiLSTM-CNN-CRF and CRF models are above 90.00%, reaching the highest F1 of 95.00% with the BERT-BiGRU-CRF model. Overall, our approach has low sentence error rates, at 8.15% with CRF and 7.11% with BiLSTM-CNNCRF taggers, and only 6.67% with BERT-BiGRU-CRF tagger.","PeriodicalId":330865,"journal":{"name":"2022 14th International Conference on Knowledge and Systems Engineering (KSE)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114485344","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Welcome Message from the KSE 2022 General Committee KSE 2022大会委员会欢迎辞
Pub Date : 2021-11-10 DOI: 10.1109/kse53942.2021.9648796
N. Trang, N. Chawla, Y. Nagai, Nguyen Thanh Thoai
The 2022 IEEE International Conference on Knowledge and Systems Engineering (KSE) is the 14th meeting of the series, held online during October 19-21, 2022 at Thai Binh Duong University. Nha Trang, Vietnam. The first KSE conference was held during October 13-17, 2009, by the College of Technology, Vietnam National University (in short, VNU), Hanoi. As an annual meeting, the past KSE conferences were held by the University of Engineering and Technology, VNU (in short, VNU-UET), Hanoi and Le Quy Don Technical University (in short, LQDTU) in 2010, by Hanoi University and VNU-UET in 2011, by Danang University of Technology, University of Da Nang, and VNU-UET in 2012, by National University of Education and VNU-UET in 2013, by VNU-UET in 2014, University of Information and Technology, VNU, Ho Chi Minh City and Japan Advanced Institute of Science and Technology (in short, JAIST) in 2015, by LQDTU, JAIST, and VNU-UET in 2016, by Hue University of Education, Hue University of Sciences, Hue University, and VNU-UET in 2017, by Telecommunications Institute of Technology, Vietnam (PTIT) and JAIST in 2018, by University of Science and Education, University of Da Nang and VNU-UET in 2019, and by College of Information and Communication Technology, Can Tho University (in short, CTU), and VNU-UET in 2020, by Artificial Intelligence Association of Thailand (AIAT), Thamasart University (TU), Sirindhorn International Insitutute of Technology, Thamasart University, National Electronics and Computer Technology Center (NECTEC), Thailand, UET-VNU, LQDTU, Vietnam, and JAIST Japan in 2021.
2022年IEEE知识与系统工程国际会议(KSE)是该系列的第14次会议,于2022年10月19日至21日在泰国平阳大学在线举行。芽庄,越南。第一届KSE会议于2009年10月13日至17日在河内的越南国立大学技术学院举行。作为年度会议,历届KSE会议分别于2010年、2011年、2012年、越南岘港理工大学、岘港理工大学、越南岘港理工大学、越南国立教育大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学、越南国立越南理工大学等举办。2015年,胡志明市和日本高等科学技术研究所(简称JAIST), 2016年,LQDTU, JAIST和VNU-UET, 2017年,顺化教育大学,顺化科技大学,顺化大学和VNU-UET, 2018年,越南电信技术学院(PTIT)和JAIST,科教大学,岘港大学和VNU-UET, 2019年,信息与通信技术学院,Can Tho大学(简称CTU),泰国人工智能协会(AIAT)、泰国泰马萨大学(TU)、诗琳通国际理工学院、泰国泰马萨大学、泰国国家电子与计算机技术中心(NECTEC)、泰国泰马萨大学、越南LQDTU和日本JAIST(2021年)。
{"title":"Welcome Message from the KSE 2022 General Committee","authors":"N. Trang, N. Chawla, Y. Nagai, Nguyen Thanh Thoai","doi":"10.1109/kse53942.2021.9648796","DOIUrl":"https://doi.org/10.1109/kse53942.2021.9648796","url":null,"abstract":"The 2022 IEEE International Conference on Knowledge and Systems Engineering (KSE) is the 14th meeting of the series, held online during October 19-21, 2022 at Thai Binh Duong University. Nha Trang, Vietnam. The first KSE conference was held during October 13-17, 2009, by the College of Technology, Vietnam National University (in short, VNU), Hanoi. As an annual meeting, the past KSE conferences were held by the University of Engineering and Technology, VNU (in short, VNU-UET), Hanoi and Le Quy Don Technical University (in short, LQDTU) in 2010, by Hanoi University and VNU-UET in 2011, by Danang University of Technology, University of Da Nang, and VNU-UET in 2012, by National University of Education and VNU-UET in 2013, by VNU-UET in 2014, University of Information and Technology, VNU, Ho Chi Minh City and Japan Advanced Institute of Science and Technology (in short, JAIST) in 2015, by LQDTU, JAIST, and VNU-UET in 2016, by Hue University of Education, Hue University of Sciences, Hue University, and VNU-UET in 2017, by Telecommunications Institute of Technology, Vietnam (PTIT) and JAIST in 2018, by University of Science and Education, University of Da Nang and VNU-UET in 2019, and by College of Information and Communication Technology, Can Tho University (in short, CTU), and VNU-UET in 2020, by Artificial Intelligence Association of Thailand (AIAT), Thamasart University (TU), Sirindhorn International Insitutute of Technology, Thamasart University, National Electronics and Computer Technology Center (NECTEC), Thailand, UET-VNU, LQDTU, Vietnam, and JAIST Japan in 2021.","PeriodicalId":330865,"journal":{"name":"2022 14th International Conference on Knowledge and Systems Engineering (KSE)","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128185590","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Keynotes 主题演讲
Pub Date : 2011-09-01 DOI: 10.1080/08870446.2011.618585
Carola Lilienthal
Today programmers do not develop applications from scratch, but they spend their time fixing, extending, modifying and enhancing existing applications. The biggest problem in their daily work is that with time maintenance mutates from structured programming to defensive programming: The code becomes too complex to be maintained. We put in code we know is stupid from an architectural point of view but it is the only solution that will hopefully work. Maintenance is more and more difficult and expensive. Our software accumulates technical debts. In this talk, you will see how you should improve your architecture and source code to prevent technical debt growing unrestricted. With the proper knowledge about well-structured architecture, refactorings for tangled code can quickly be found. Complex code can be eliminated, and maintenance costs will be reduced. Bio Carola Lilienthal studied computer science at the University of Hamburg from 1988 to 1995, and in 2008 she received her doctoral degree in computer science at the University of Hamburg (Supervising Professors: Christiane Floyd and Claus Lewerentz). Today, Dr. Carola Lilienthal is managing director of WPS Workplace Solutions GmbH and is responsible for the department of software architecture. Since 2003, Dr. Carola Lilienthal has been analyzing architecture in Java, C #, C ++, ABAP and PHP throughout Germany, and advising development teams on how to improve the longevity of their software systems. In 2015, she summarized her experiences from over a hundred analyzes in the book Long-living software architectures. She is particularly interested in the education of software architects, which is why she is an active member of iSAQB, the International Software Architecture Quality Board e.V., and regularly disseminates her knowledge at conferences, in articles and training courses.
今天,程序员不再从头开始开发应用程序,而是把时间花在修复、扩展、修改和增强现有应用程序上。他们日常工作中最大的问题是,随着时间的推移,维护从结构化编程变成了防御性编程:代码变得太复杂而无法维护。从架构的角度来看,我们知道代码是愚蠢的,但这是唯一可行的解决方案。维修越来越困难,费用也越来越高。我们的软件积累了技术债务。在本次演讲中,您将看到应该如何改进架构和源代码,以防止技术债务无限制地增长。有了关于结构良好的体系结构的适当知识,就可以很快地找到复杂代码的重构。可以消除复杂的代码,并降低维护成本。Carola Lilienthal于1988年至1995年在汉堡大学学习计算机科学,并于2008年获得汉堡大学计算机科学博士学位(指导教授:Christiane Floyd和Claus Lewerentz)。Carola Lilienthal博士是WPS Workplace Solutions GmbH的董事总经理,负责软件架构部门。自2003年以来,Carola Lilienthal博士一直在德国各地分析Java、c#、c++、ABAP和PHP的架构,并就如何提高软件系统的寿命为开发团队提供建议。2015年,她在《长寿的软件架构》一书中总结了她从一百多个分析中获得的经验。她对软件架构师的教育特别感兴趣,这就是为什么她是iSAQB(国际软件架构质量委员会)的活跃成员,并定期在会议、文章和培训课程中传播她的知识。
{"title":"Keynotes","authors":"Carola Lilienthal","doi":"10.1080/08870446.2011.618585","DOIUrl":"https://doi.org/10.1080/08870446.2011.618585","url":null,"abstract":"Today programmers do not develop applications from scratch, but they spend their time fixing, extending, modifying and enhancing existing applications. The biggest problem in their daily work is that with time maintenance mutates from structured programming to defensive programming: The code becomes too complex to be maintained. We put in code we know is stupid from an architectural point of view but it is the only solution that will hopefully work. Maintenance is more and more difficult and expensive. Our software accumulates technical debts. In this talk, you will see how you should improve your architecture and source code to prevent technical debt growing unrestricted. With the proper knowledge about well-structured architecture, refactorings for tangled code can quickly be found. Complex code can be eliminated, and maintenance costs will be reduced. Bio Carola Lilienthal studied computer science at the University of Hamburg from 1988 to 1995, and in 2008 she received her doctoral degree in computer science at the University of Hamburg (Supervising Professors: Christiane Floyd and Claus Lewerentz). Today, Dr. Carola Lilienthal is managing director of WPS Workplace Solutions GmbH and is responsible for the department of software architecture. Since 2003, Dr. Carola Lilienthal has been analyzing architecture in Java, C #, C ++, ABAP and PHP throughout Germany, and advising development teams on how to improve the longevity of their software systems. In 2015, she summarized her experiences from over a hundred analyzes in the book Long-living software architectures. She is particularly interested in the education of software architects, which is why she is an active member of iSAQB, the International Software Architecture Quality Board e.V., and regularly disseminates her knowledge at conferences, in articles and training courses.","PeriodicalId":330865,"journal":{"name":"2022 14th International Conference on Knowledge and Systems Engineering (KSE)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129346157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2022 14th International Conference on Knowledge and Systems Engineering (KSE)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1