首页 > 最新文献

2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)最新文献

英文 中文
Evaluate different machine learning techniques for classifying sleep stages on single-channel EEG 评估不同机器学习技术在单通道脑电图上的睡眠阶段分类
Shahnawaz Qureshi, S. Vanichayobon
In this paper, we propose 3 different machine learning techniques such as Random Forest, Bagging and Support Vector Machine along with time domain feature for classifying sleep stages based on single-channel EEG. Whole-night polysomnograms from 25 subjects were recorded employing R&K standard. The evolved process investigated the EEG signals of (C4-A1) for sleep staging. Automatic and manual scoring results were associated on an epoch-by-epoch basis. An entire 96,000 data samples 30s sleep EEG epoch were calculated and applied for performance evaluation. The epoch-by-epoch assessment was created by classifying the EEG epochs into six stages (W/S1/S2/S3/S4/REM) according to proposed method and manual scoring. Result shows that Random Forest classifiers achieve the overall accuracy; specificity and sensitivity level of 97.73%, 96.3% and 99.51% respectively.
在本文中,我们提出了3种不同的机器学习技术,如随机森林、Bagging和支持向量机,以及时域特征,用于基于单通道EEG的睡眠阶段分类。采用R&K标准记录25名受试者的夜间多导睡眠图。进化过程研究(C4-A1)脑电信号对睡眠分期的影响。自动和手动评分结果在一个epoch-by-epoch的基础上相关联。计算了96000个数据样本30秒睡眠脑电历元并应用于性能评价。根据提出的方法和人工评分方法,将脑电分期分为W/S1/S2/S3/S4/REM 6个阶段,形成逐期评价。结果表明,随机森林分类器达到了整体准确率;特异性97.73%,敏感性96.3%,敏感性99.51%。
{"title":"Evaluate different machine learning techniques for classifying sleep stages on single-channel EEG","authors":"Shahnawaz Qureshi, S. Vanichayobon","doi":"10.1109/JCSSE.2017.8025949","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025949","url":null,"abstract":"In this paper, we propose 3 different machine learning techniques such as Random Forest, Bagging and Support Vector Machine along with time domain feature for classifying sleep stages based on single-channel EEG. Whole-night polysomnograms from 25 subjects were recorded employing R&K standard. The evolved process investigated the EEG signals of (C4-A1) for sleep staging. Automatic and manual scoring results were associated on an epoch-by-epoch basis. An entire 96,000 data samples 30s sleep EEG epoch were calculated and applied for performance evaluation. The epoch-by-epoch assessment was created by classifying the EEG epochs into six stages (W/S1/S2/S3/S4/REM) according to proposed method and manual scoring. Result shows that Random Forest classifiers achieve the overall accuracy; specificity and sensitivity level of 97.73%, 96.3% and 99.51% respectively.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"28 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82759120","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Modeling realistic virtual pulse of radial artery pressure waveform using haptic interface 基于触觉界面的真实虚拟脉冲桡动脉压力波形建模
Moragot Kandee, P. Boonbrahm, Valla Tantayotai
This paper shows an investigation of the ability of using various waveform generated by mathematical model on haptic device. Realistic virtual pulse measurement and diagnostic can be done using haptic device with pulse generated waveform, Augmented Reality (AR) environment and mannequin. The aim of this work is to propose a mathematical model for generating pulse pattern in different type of abnormal pulse waves and test them on the Phantom Omni device under AR environment. The radial arterial waveforms were generated by the setting of pulse parameters and superimposed sine waves to make the new waveforms representing various diseases. The system can simulate the radial arterial pulse waves of some diseases. This modeling technique can be used in training the nursing or health sciences students on the ability to classify various type of diseases that related to the pulse waveform.
本文研究了利用数学模型产生的各种波形在触觉设备上的应用能力。利用具有脉冲产生波形的触觉设备、增强现实环境和人体模型,可以实现真实的虚拟脉搏测量和诊断。本文旨在建立不同类型异常脉冲波产生脉冲模式的数学模型,并在增强现实环境下的Phantom Omni设备上进行测试。通过设定脉搏参数,叠加正弦波,生成代表各种疾病的新波形。该系统可以模拟一些疾病的桡动脉脉搏波。这种建模技术可用于培训护理或健康科学专业的学生对与脉搏波形相关的各种疾病进行分类的能力。
{"title":"Modeling realistic virtual pulse of radial artery pressure waveform using haptic interface","authors":"Moragot Kandee, P. Boonbrahm, Valla Tantayotai","doi":"10.1109/JCSSE.2017.8025954","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025954","url":null,"abstract":"This paper shows an investigation of the ability of using various waveform generated by mathematical model on haptic device. Realistic virtual pulse measurement and diagnostic can be done using haptic device with pulse generated waveform, Augmented Reality (AR) environment and mannequin. The aim of this work is to propose a mathematical model for generating pulse pattern in different type of abnormal pulse waves and test them on the Phantom Omni device under AR environment. The radial arterial waveforms were generated by the setting of pulse parameters and superimposed sine waves to make the new waveforms representing various diseases. The system can simulate the radial arterial pulse waves of some diseases. This modeling technique can be used in training the nursing or health sciences students on the ability to classify various type of diseases that related to the pulse waveform.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"115 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79344052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Automatic microaneurysm detection using Multi-level Threshold based on ISODATA 基于ISODATA的多级阈值自动微动脉瘤检测
Tanin Intaramanee, Ratanak Khoeun, K. Chinnasarn
Diabetic Retinopathy is one of the most serious diseases that can lead to blindness. The small swelling blood regions or microaneurysms are the early sign of Diabetic Retinopathy. Detecting that a patient has got the Diabetic Retinopathy at the earliest stage as possible can help to prevent him/her from the vision lost. However, automatic microaneurysm detection is still a challenging topic for medical image processing researchers. This is because of the varieties of microaneurysm characteristic such as size, contrast, shape, and data distribution. In this paper, we propose an approach to automatically detect microaneurysms using Multi-level Threshold based on ISODATA. The proposed method consists of two main steps: 1) preprocessing and 2) feature extraction. In the preprocessing step, Contrast Limited Adaptive Histogram Equalization, Gaussian Filter and Median Filter are applied to enhance the image quality. Next, in the feature extraction step, Multi-level Threshold based on ISODATA and Noise Removing Techniques are adopted to remove non-microaneurysm objects. The 89 retinal fundus images from a public database DIARETDB1 are used as a dataset. By comparing with the ground truth, the proposed approach provides the reasonable results with sensitivity of 62.82%, specificity of 93.60% and accuracy of 93.43%.
糖尿病视网膜病变是可导致失明的最严重疾病之一。小的肿胀血区或微动脉瘤是糖尿病视网膜病变的早期征兆。尽早发现患者是否患有糖尿病视网膜病变,可以帮助患者避免视力丧失。然而,对于医学图像处理研究者来说,微动脉瘤的自动检测仍然是一个具有挑战性的课题。这是由于微动脉瘤的各种特征,如大小、对比度、形状和数据分布。本文提出了一种基于ISODATA的多级阈值自动检测微动脉瘤的方法。该方法主要包括两个步骤:1)预处理和2)特征提取。在预处理步骤中,采用对比度有限自适应直方图均衡化、高斯滤波和中值滤波来提高图像质量。接下来,在特征提取步骤中,采用基于ISODATA的多级阈值和去噪技术去除非微动脉瘤目标。使用来自公共数据库DIARETDB1的89张视网膜眼底图像作为数据集。通过与地面真实值的比较,该方法的灵敏度为62.82%,特异度为93.60%,准确率为93.43%。
{"title":"Automatic microaneurysm detection using Multi-level Threshold based on ISODATA","authors":"Tanin Intaramanee, Ratanak Khoeun, K. Chinnasarn","doi":"10.1109/JCSSE.2017.8025958","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025958","url":null,"abstract":"Diabetic Retinopathy is one of the most serious diseases that can lead to blindness. The small swelling blood regions or microaneurysms are the early sign of Diabetic Retinopathy. Detecting that a patient has got the Diabetic Retinopathy at the earliest stage as possible can help to prevent him/her from the vision lost. However, automatic microaneurysm detection is still a challenging topic for medical image processing researchers. This is because of the varieties of microaneurysm characteristic such as size, contrast, shape, and data distribution. In this paper, we propose an approach to automatically detect microaneurysms using Multi-level Threshold based on ISODATA. The proposed method consists of two main steps: 1) preprocessing and 2) feature extraction. In the preprocessing step, Contrast Limited Adaptive Histogram Equalization, Gaussian Filter and Median Filter are applied to enhance the image quality. Next, in the feature extraction step, Multi-level Threshold based on ISODATA and Noise Removing Techniques are adopted to remove non-microaneurysm objects. The 89 retinal fundus images from a public database DIARETDB1 are used as a dataset. By comparing with the ground truth, the proposed approach provides the reasonable results with sensitivity of 62.82%, specificity of 93.60% and accuracy of 93.43%.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"144 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77407455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A semantic approach for question answering using DBpedia and WordNet 使用DBpedia和WordNet进行问题回答的语义方法
Kittiphong Sengloiluean, N. Arch-int, S. Arch-int, Theerayut Thongkrau
Semantic Question Answering (SQA) was concerned about the natural language processing. The purpose of this study was to help facilitate the users to access the information through the natural language and to obtain the concise and needed information. As considered the current studies, it was found that this processing still encountered the problems of flexibility and accuracy, particularly those of the question processing, which was a very important processing for developing question answering system. Thus, this study proposed a semantic approach for question answering using DBpedia and WordNet. For this paper, the techniques for solving the problems were proposed consisting of (1) extracting named entities from the question and solving the problems of similarities of named entities, (2) extracting properties from the question and solving the problems of similarities of properties, and (3) evaluating the accurate capability of the answer of question. This approach evaluated the test dataset from TREC question collections, DBpedia and achieved an F-measure score of 93.43%, an average precision of 92.73%, and an average recall of 94.15% over 500 questions.
语义问答(Semantic Question answer, SQA)是一个关注自然语言处理的问题。本研究的目的是帮助用户方便地通过自然语言获取信息,并获得简洁和所需的信息。考虑到目前的研究,我们发现这种处理仍然遇到了灵活性和准确性的问题,特别是问题处理的问题,这是开发问答系统的一个非常重要的处理。因此,本研究提出了一种使用DBpedia和WordNet进行语义问答的方法。本文提出了从问题中提取命名实体并求解命名实体的相似度问题,从问题中提取属性并求解属性的相似度问题,以及评价问题答案的准确性的技术。该方法对来自TREC问题集、DBpedia的测试数据集进行了评估,500个问题的F-measure得分为93.43%,平均准确率为92.73%,平均召回率为94.15%。
{"title":"A semantic approach for question answering using DBpedia and WordNet","authors":"Kittiphong Sengloiluean, N. Arch-int, S. Arch-int, Theerayut Thongkrau","doi":"10.1109/JCSSE.2017.8025918","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025918","url":null,"abstract":"Semantic Question Answering (SQA) was concerned about the natural language processing. The purpose of this study was to help facilitate the users to access the information through the natural language and to obtain the concise and needed information. As considered the current studies, it was found that this processing still encountered the problems of flexibility and accuracy, particularly those of the question processing, which was a very important processing for developing question answering system. Thus, this study proposed a semantic approach for question answering using DBpedia and WordNet. For this paper, the techniques for solving the problems were proposed consisting of (1) extracting named entities from the question and solving the problems of similarities of named entities, (2) extracting properties from the question and solving the problems of similarities of properties, and (3) evaluating the accurate capability of the answer of question. This approach evaluated the test dataset from TREC question collections, DBpedia and achieved an F-measure score of 93.43%, an average precision of 92.73%, and an average recall of 94.15% over 500 questions.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"10 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89119296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A robust algorithm for R peak detection based on optimal Discrete Wavelet Transform 基于最优离散小波变换的鲁棒R峰值检测算法
Anurak Thungtong
Automated ECG signal processing can assist in diagnosing several heart diseases. Many R peak detection methods have been studied because the accuracy of R peak detection significantly affects the quality of subsequent ECG feature extraction. Two important steps in R peak detection algorithm that draw attention over researchers are the preprocessing and thresholding stages. Among several methods, wavelet transform is a widely used method for removing noise in the preprocessing stage. Various proposed algorithms require prior knowledge of frequency spectrum of the signal under consideration in order to select the wavelet detail coefficients in the reconstruction process. Moreover, parameter fine tuning is generally involved in threshold selection to accomplish high detection accuracy. As a result, it may be difficult to utilize these methods for general ECG data sets. Accordingly, we propose an automatic and parameter free method that optimally selects the appropriate detail components for wavelet reconstruction as well as the adaptive threshold. The proposed algorithm employs the analysis of probability density function of the processed ECG signal. The validation of the algorithm was performed over the MIT-BIH database and has produced an average sensitivity of 99.63% and specificity of 99.78% which is in the same range as the previously proposed approaches.
自动心电信号处理可以帮助诊断多种心脏疾病。由于R峰检测的准确性对后续心电特征提取的质量影响很大,因此人们研究了许多R峰检测方法。在R峰值检测算法中,预处理和阈值处理是研究人员关注的两个重要步骤。在各种方法中,小波变换是一种广泛应用于预处理阶段的去噪方法。为了在重建过程中选择小波细节系数,各种提出的算法都需要对所考虑的信号的频谱有先验知识。此外,阈值的选择通常涉及参数微调,以达到较高的检测精度。因此,很难将这些方法应用于一般的心电数据集。因此,我们提出了一种自动的、无参数的方法,以最优地选择合适的细节分量进行小波重构和自适应阈值。该算法对处理后的心电信号进行概率密度函数分析。该算法在MIT-BIH数据库上进行了验证,产生了99.63%的平均灵敏度和99.78%的特异性,与之前提出的方法在相同的范围内。
{"title":"A robust algorithm for R peak detection based on optimal Discrete Wavelet Transform","authors":"Anurak Thungtong","doi":"10.1109/JCSSE.2017.8025931","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025931","url":null,"abstract":"Automated ECG signal processing can assist in diagnosing several heart diseases. Many R peak detection methods have been studied because the accuracy of R peak detection significantly affects the quality of subsequent ECG feature extraction. Two important steps in R peak detection algorithm that draw attention over researchers are the preprocessing and thresholding stages. Among several methods, wavelet transform is a widely used method for removing noise in the preprocessing stage. Various proposed algorithms require prior knowledge of frequency spectrum of the signal under consideration in order to select the wavelet detail coefficients in the reconstruction process. Moreover, parameter fine tuning is generally involved in threshold selection to accomplish high detection accuracy. As a result, it may be difficult to utilize these methods for general ECG data sets. Accordingly, we propose an automatic and parameter free method that optimally selects the appropriate detail components for wavelet reconstruction as well as the adaptive threshold. The proposed algorithm employs the analysis of probability density function of the processed ECG signal. The validation of the algorithm was performed over the MIT-BIH database and has produced an average sensitivity of 99.63% and specificity of 99.78% which is in the same range as the previously proposed approaches.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"18 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73620870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
The classification of sets of medical procedures used in the treatment of Diabetes and/or Hypertension 用于治疗糖尿病和/或高血压的医疗程序的分类
Kanokwan Rungreangsuparat, S. Kitisin, K. Sripanidkulchai
The advancement of technology to support data storage is easy to store with large volumes of data. In order to make data storage to be used extensively, the standard of data storage is formed. The World Health Organization defines numbers of the standard medical procedures to cover the all treatments without classifying any medical procedures by diseases. The selection of medical procedures is based on a patient's symptoms. Therefore, if the sets of medical procedures can identified, we may know the diseases of the patient or it can be used in disease surveillance. In addition, diabetes and hypertension are silent killers that have been threatening numbers of Thai people and also lead to many serious diseases. This research identified sets of medical procedures related to diabetes and/or hypertension using C4.5 and Naive Bayes algorithms. The results showed that C4.5 could identify sets of medical procedures related to Diabetes and/or Hypertension more effectively than the Naive Bayes algorithm.
支持数据存储的技术的进步使得存储大量数据变得容易。为了使数据存储得到广泛应用,形成了数据存储标准。世界卫生组织定义了标准医疗程序的数量,以涵盖所有治疗,而不按疾病分类任何医疗程序。医疗程序的选择是基于病人的症状。因此,如果这套医疗程序可以识别,我们就可以知道患者的疾病情况,或者可以用于疾病监测。此外,糖尿病和高血压是无声的杀手,一直威胁着许多泰国人,也导致许多严重的疾病。本研究使用C4.5和朴素贝叶斯算法确定了与糖尿病和/或高血压相关的医疗程序集。结果表明,C4.5可以比朴素贝叶斯算法更有效地识别与糖尿病和/或高血压相关的医疗程序集。
{"title":"The classification of sets of medical procedures used in the treatment of Diabetes and/or Hypertension","authors":"Kanokwan Rungreangsuparat, S. Kitisin, K. Sripanidkulchai","doi":"10.1109/JCSSE.2017.8025939","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025939","url":null,"abstract":"The advancement of technology to support data storage is easy to store with large volumes of data. In order to make data storage to be used extensively, the standard of data storage is formed. The World Health Organization defines numbers of the standard medical procedures to cover the all treatments without classifying any medical procedures by diseases. The selection of medical procedures is based on a patient's symptoms. Therefore, if the sets of medical procedures can identified, we may know the diseases of the patient or it can be used in disease surveillance. In addition, diabetes and hypertension are silent killers that have been threatening numbers of Thai people and also lead to many serious diseases. This research identified sets of medical procedures related to diabetes and/or hypertension using C4.5 and Naive Bayes algorithms. The results showed that C4.5 could identify sets of medical procedures related to Diabetes and/or Hypertension more effectively than the Naive Bayes algorithm.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"68 1","pages":"1-7"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76290406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Extracting UML class diagrams from software requirements in Thai using NLP 使用NLP从泰语软件需求中提取UML类图
Mathawan Jaiwai, Usa Sammapun
In software development, requirements, normally written in natural language, are documents that specify what users want in software products. Software developers then analyze these requirements to create domain models represented in UML diagrams in an attempt to comprehend what users need in the software products. These domain models are usually converted into design models and finally carried over into classes in source code. Thus, domain models have an impact on the final software products. However, creating correct domain models can be difficult when software developers are not skilled. Moreover, even for skilled developers, when requirements are large, wading through all requirements to create domain models can take times and might result in errors. Therefore, researchers have studied various approaches to apply natural language processing techniques to transform requirements written in natural language into UML diagrams. Those researches focus on requirements written in English. This paper proposes an approach to process requirements written in Thai to extract UML class diagrams using natural language processing techniques. The UML class diagram extraction is based on transformation rules that identify classes and attributes from requirements. The results are evaluated with recall and precision using truth values created by humans. Future works include identifying operations and relationships from requirements to complete class diagram extraction. Our research should benefit Thai software developers by reducing time in requirement analysis and also helping novice software developers to create correct domain models represented in UML class diagram.
在软件开发中,需求通常用自然语言编写,是指定用户在软件产品中需要什么的文档。然后,软件开发人员分析这些需求,以创建用UML图表示的领域模型,试图理解用户在软件产品中需要什么。这些领域模型通常被转换为设计模型,并最终被转移到源代码中的类中。因此,领域模型对最终的软件产品有影响。然而,当软件开发人员不熟练时,创建正确的领域模型是很困难的。此外,即使是熟练的开发人员,当需求很大时,费力地浏览所有需求来创建领域模型可能会花费时间,并可能导致错误。因此,研究人员已经研究了各种方法来应用自然语言处理技术,将用自然语言编写的需求转换为UML图。这些研究的重点是用英语写的需求。本文提出了一种处理用泰语编写的需求的方法,使用自然语言处理技术提取UML类图。UML类图的提取是基于从需求中识别类和属性的转换规则。使用人类创建的真值来评估结果的召回率和精度。未来的工作包括识别操作和关系,从需求到完整的类图提取。我们的研究应该通过减少需求分析的时间和帮助新手软件开发人员创建用UML类图表示的正确的领域模型而使泰国的软件开发人员受益。
{"title":"Extracting UML class diagrams from software requirements in Thai using NLP","authors":"Mathawan Jaiwai, Usa Sammapun","doi":"10.1109/JCSSE.2017.8025938","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025938","url":null,"abstract":"In software development, requirements, normally written in natural language, are documents that specify what users want in software products. Software developers then analyze these requirements to create domain models represented in UML diagrams in an attempt to comprehend what users need in the software products. These domain models are usually converted into design models and finally carried over into classes in source code. Thus, domain models have an impact on the final software products. However, creating correct domain models can be difficult when software developers are not skilled. Moreover, even for skilled developers, when requirements are large, wading through all requirements to create domain models can take times and might result in errors. Therefore, researchers have studied various approaches to apply natural language processing techniques to transform requirements written in natural language into UML diagrams. Those researches focus on requirements written in English. This paper proposes an approach to process requirements written in Thai to extract UML class diagrams using natural language processing techniques. The UML class diagram extraction is based on transformation rules that identify classes and attributes from requirements. The results are evaluated with recall and precision using truth values created by humans. Future works include identifying operations and relationships from requirements to complete class diagram extraction. Our research should benefit Thai software developers by reducing time in requirement analysis and also helping novice software developers to create correct domain models represented in UML class diagram.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"15 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89596192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Vertebral pose segmentation on low radiation image using Convergence Gravity Force 基于汇聚重力的低辐射图像椎体位姿分割
Jakapong Boonyai, Suwanna Rasmequan
Vertebral pose segmentation is an important factor in diagnosing diseases such as osteoporosis, osteopenia and scoliosis. Low radiation X-ray images are often used to diagnose such diseases. This has been done to reduce patients risk exposure of over dose radiation which may cause from a series of treatments. In this respect, it led to a low accuracy in vertebral pose detection. In this paper, we proposed to improve the automate segmentation of low quality image of vertebral pose with a more generalized technique. In the proposed method, there are three main steps. Firstly, in the pre-processing step, Auto Cropped, Multi-Threshold and Canny Edge Detection are applied to find the vertebral bone structure from the original image. Secondly, Feature Analysis and Gravity Force were used to find the region of interest or the area of each pose. Finally, Colormaps, Intensity Diagnosis and Angle Analysis are adopted to segment each vertebral pose from candidate areas retrieved from second step. The experimental results which were compared with ground truth shown that the proposed approach can estimate vertebral pose with Precision at 79.61% and Recall at 77.11%.
椎体位姿分割是诊断骨质疏松、骨质减少和脊柱侧凸等疾病的重要因素。低辐射x射线图像常用于诊断此类疾病。这样做是为了减少患者暴露于一系列治疗可能引起的过量辐射的风险。在这方面,它导致椎体姿态检测精度低。在本文中,我们提出了一种更广义的技术来改进低质量椎体姿态图像的自动分割。在提出的方法中,主要有三个步骤。首先,在预处理步骤中,采用自动裁剪、多阈值和Canny边缘检测从原始图像中找到椎体骨结构;其次,利用特征分析和重力法找到感兴趣的区域或每个姿态的面积;最后,采用colormap、Intensity Diagnosis和Angle Analysis从第二步提取的候选区域中分割出每个椎体姿态。实验结果表明,该方法对椎体姿态的估计精度为79.61%,查全率为77.11%。
{"title":"Vertebral pose segmentation on low radiation image using Convergence Gravity Force","authors":"Jakapong Boonyai, Suwanna Rasmequan","doi":"10.1109/JCSSE.2017.8025959","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025959","url":null,"abstract":"Vertebral pose segmentation is an important factor in diagnosing diseases such as osteoporosis, osteopenia and scoliosis. Low radiation X-ray images are often used to diagnose such diseases. This has been done to reduce patients risk exposure of over dose radiation which may cause from a series of treatments. In this respect, it led to a low accuracy in vertebral pose detection. In this paper, we proposed to improve the automate segmentation of low quality image of vertebral pose with a more generalized technique. In the proposed method, there are three main steps. Firstly, in the pre-processing step, Auto Cropped, Multi-Threshold and Canny Edge Detection are applied to find the vertebral bone structure from the original image. Secondly, Feature Analysis and Gravity Force were used to find the region of interest or the area of each pose. Finally, Colormaps, Intensity Diagnosis and Angle Analysis are adopted to segment each vertebral pose from candidate areas retrieved from second step. The experimental results which were compared with ground truth shown that the proposed approach can estimate vertebral pose with Precision at 79.61% and Recall at 77.11%.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"16 1","pages":"1-6"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87688592","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A performance comparison of Apache Tez and MapReduce with data compression on Hadoop cluster Apache Tez和MapReduce在Hadoop集群上数据压缩的性能比较
Kritwara Rattanaopas
Big data is a popular topic on cloud computing research. The main characteristics of big data are volume, velocity and variety. These characteristics are difficult to handle by using traditional softwares and methods. Hadoop is open-source framework software which was developed to provide solutions for handling several domains of big data problems. For big data analytic, MapReduce framework is a main engine of Hadoop cluster and widely used nowadays. It uses a batch oriented processing. Apache also developed an alternative engine called “Tez”. It supports an interactive query and does not write temporary data into HDFS. In this paper, we focus on the performance comparison between MapReduce and Tez. We also investigate the performance of these two engines with the compression of input files and map output files. Bzip is a compression algorithm used for input files and snappy is used for map output files. Word-count and terasort benchmarks are used in our experiments. For the word-count benchmark, the results show that Tez engine always has better execution-time than MapReduce engine for both of compressed data or non-compressed data. It can reduce an execution-time up to 39% comparing with the execution time of MapReduce engine. In contrast, the results show that Tez engine usually has higher execution-time than MapReduce engine up to 13% for terasort benchmark. The results also show that the performance of compressing map output files with snappy provides better performance on execution time for both benchmarks.
大数据是云计算研究的热门话题。大数据的主要特点是量大、速度快、种类多。这些特点是传统的软件和方法难以处理的。Hadoop是开源框架软件,它的开发是为了提供解决方案来处理几个领域的大数据问题。对于大数据分析,MapReduce框架是Hadoop集群的主要引擎,目前应用广泛。它使用面向批处理的处理。Apache还开发了一种名为“Tez”的替代引擎。它支持交互式查询,不将临时数据写入HDFS。在本文中,我们着重于MapReduce和Tez之间的性能比较。我们还研究了这两个引擎在压缩输入文件和映射输出文件方面的性能。Bzip是用于输入文件的压缩算法,snappy用于映射输出文件。在我们的实验中使用了单词计数和分类基准。对于单词计数的基准测试,结果表明Tez引擎无论对压缩数据还是非压缩数据都比MapReduce引擎有更好的执行时间。与MapReduce引擎相比,它可以减少高达39%的执行时间。相比之下,结果表明Tez引擎通常比MapReduce引擎具有更高的执行时间,在terassort基准测试中高达13%。结果还表明,使用snappy压缩映射输出文件的性能为两个基准测试提供了更好的执行时间性能。
{"title":"A performance comparison of Apache Tez and MapReduce with data compression on Hadoop cluster","authors":"Kritwara Rattanaopas","doi":"10.1109/JCSSE.2017.8025950","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025950","url":null,"abstract":"Big data is a popular topic on cloud computing research. The main characteristics of big data are volume, velocity and variety. These characteristics are difficult to handle by using traditional softwares and methods. Hadoop is open-source framework software which was developed to provide solutions for handling several domains of big data problems. For big data analytic, MapReduce framework is a main engine of Hadoop cluster and widely used nowadays. It uses a batch oriented processing. Apache also developed an alternative engine called “Tez”. It supports an interactive query and does not write temporary data into HDFS. In this paper, we focus on the performance comparison between MapReduce and Tez. We also investigate the performance of these two engines with the compression of input files and map output files. Bzip is a compression algorithm used for input files and snappy is used for map output files. Word-count and terasort benchmarks are used in our experiments. For the word-count benchmark, the results show that Tez engine always has better execution-time than MapReduce engine for both of compressed data or non-compressed data. It can reduce an execution-time up to 39% comparing with the execution time of MapReduce engine. In contrast, the results show that Tez engine usually has higher execution-time than MapReduce engine up to 13% for terasort benchmark. The results also show that the performance of compressing map output files with snappy provides better performance on execution time for both benchmarks.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"42 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90120331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Strabismus screening by Eye Tracker and games 眼动仪和游戏的斜视筛查
Udomchai Saisara, P. Boonbrahm, Achara Chaiwiriya
More than 5% of Thai people have strabismus. Strabismus is known as cross-eyed or wall-eyed because the visual field angle of two eyes is not parallel. The amblyopia disease is the cause of strabismus in kids. Strabismus can be completely cured if the strabismus screening can be made in early stage. Currently, strabismus screening includes methods such as Hirschberg test, cover test and Krimsky test, and etc. The strabismus screening in kids is difficult and takes a lot time in special room. This research intend to develop a computer system to assist strabismus screening using the combination of computer games and eye tracking devices so that the screening results will be more accurate and exact. This screening technique requires shorter time and it is easy to use, so it is better in terms of efficiency and reducing time for strabismus screening.
超过5%的泰国人患有斜视。斜视被称为对眼或壁眼,因为两只眼睛的视野角度不平行。弱视疾病是造成儿童斜视的原因。如果能及早进行斜视筛查,斜视是可以完全治愈的。目前,斜视的筛查方法主要有Hirschberg试验、cover试验、Krimsky试验等。儿童斜视的筛查是一项困难的工作,需要在特殊的房间里花费大量的时间。本研究拟开发电脑游戏与眼动追踪设备相结合的辅助斜视筛查的电脑系统,使筛查结果更加准确、准确。这种筛查技术所需时间短,使用方便,因此在斜视筛查的效率和减少时间方面较好。
{"title":"Strabismus screening by Eye Tracker and games","authors":"Udomchai Saisara, P. Boonbrahm, Achara Chaiwiriya","doi":"10.1109/JCSSE.2017.8025956","DOIUrl":"https://doi.org/10.1109/JCSSE.2017.8025956","url":null,"abstract":"More than 5% of Thai people have strabismus. Strabismus is known as cross-eyed or wall-eyed because the visual field angle of two eyes is not parallel. The amblyopia disease is the cause of strabismus in kids. Strabismus can be completely cured if the strabismus screening can be made in early stage. Currently, strabismus screening includes methods such as Hirschberg test, cover test and Krimsky test, and etc. The strabismus screening in kids is difficult and takes a lot time in special room. This research intend to develop a computer system to assist strabismus screening using the combination of computer games and eye tracking devices so that the screening results will be more accurate and exact. This screening technique requires shorter time and it is easy to use, so it is better in terms of efficiency and reducing time for strabismus screening.","PeriodicalId":6460,"journal":{"name":"2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)","volume":"14 1","pages":"1-5"},"PeriodicalIF":0.0,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82921637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
2017 14th International Joint Conference on Computer Science and Software Engineering (JCSSE)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1