Pub Date : 2022-03-09DOI: 10.1142/s2196888822500178
Minh Duc Nguyen, Giang H. Nguyen-Thi, Cuong H. Nguyen-Dinh
Legal ontologies play a key role in various legal applications and have been broadly used by many stakeholders. Innovative systems and ontologies in the law hold potential to conduct legal research. With the needs for legal information management in smart applications, especially for Vietnamese law, it is vitally important to construct core legal ontologies for knowledge representation. This study proposes a core ontology for Vietnamese legal documents which covers general legal domain called as ViLO. The ViLO ontology mainly consists of related institutions of Vietnamese political system, types and structures of legal documents. The method of the NeOn-based collaborations among domain experts and ontology engineers was conducted to build up the ViLO ontology. Through FOCA-based validation results, the proposed method was shown to be effective and efficient. The resulting ontology was demonstrated to be reliable and enriched. The ViLO ontology is supposed to be a basis for further constructions of domain ontologies and artificial intelligence applications in Vietnamese law.
{"title":"ViLO - A Core Ontology of Vietnamese Legal Documents","authors":"Minh Duc Nguyen, Giang H. Nguyen-Thi, Cuong H. Nguyen-Dinh","doi":"10.1142/s2196888822500178","DOIUrl":"https://doi.org/10.1142/s2196888822500178","url":null,"abstract":"Legal ontologies play a key role in various legal applications and have been broadly used by many stakeholders. Innovative systems and ontologies in the law hold potential to conduct legal research. With the needs for legal information management in smart applications, especially for Vietnamese law, it is vitally important to construct core legal ontologies for knowledge representation. This study proposes a core ontology for Vietnamese legal documents which covers general legal domain called as ViLO. The ViLO ontology mainly consists of related institutions of Vietnamese political system, types and structures of legal documents. The method of the NeOn-based collaborations among domain experts and ontology engineers was conducted to build up the ViLO ontology. Through FOCA-based validation results, the proposed method was shown to be effective and efficient. The resulting ontology was demonstrated to be reliable and enriched. The ViLO ontology is supposed to be a basis for further constructions of domain ontologies and artificial intelligence applications in Vietnamese law.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"2142 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130006124","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-02-14DOI: 10.1142/s2196888822500166
A. Kasperski, P. Zieliński
In this paper, a production planning problem with inventory and backordering levels is discussed. It is assumed that cumulative demands in periods are uncertain and an interval uncertainty representation with continuous budget is used to model this uncertainty. The robust minmax criterion is applied to compute an optimal production plan. A row and column generation algorithm is constructed for solving the problem. Results of some computational tests are shown which demonstrate that the algorithm is efficient for the instances with up to 100 periods and returns solutions that are robust against the uncertainty in demands.
{"title":"Solving Robust Production Planning Problem with Interval Budgeted Uncertainty in Cumulative Demands","authors":"A. Kasperski, P. Zieliński","doi":"10.1142/s2196888822500166","DOIUrl":"https://doi.org/10.1142/s2196888822500166","url":null,"abstract":"In this paper, a production planning problem with inventory and backordering levels is discussed. It is assumed that cumulative demands in periods are uncertain and an interval uncertainty representation with continuous budget is used to model this uncertainty. The robust minmax criterion is applied to compute an optimal production plan. A row and column generation algorithm is constructed for solving the problem. Results of some computational tests are shown which demonstrate that the algorithm is efficient for the instances with up to 100 periods and returns solutions that are robust against the uncertainty in demands.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-02-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115407604","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-24DOI: 10.1142/s2196888822500142
R. Arora, Arvinder Kaur
Software Fault Prediction (SFP) is the most persuasive research area of software engineering. Software Fault Prediction which is carried out within the same software project is known as With-In Fault Prediction. However, local data repositories are not enough to build the model of With-in software Fault prediction. The idea of cross-project fault prediction (CPFP) has been suggested in recent years, which aims to construct a prediction model on one project, and use that model to predict the other project. However, CPFP requires that both the training and testing datasets use the same set of metrics. As a consequence, traditional CPFP approaches are challenging to implement through projects with diverse metric sets. The specific case of CPFP is Heterogeneous Fault Prediction (HFP), which allows the program to predict faults among projects with diverse metrics. The proposed framework aims to achieve an HFP model by implementing Feature Selection on both the source and target datasets to build an efficient prediction model using supervised machine learning techniques. Our approach is applied on two open-source projects, Linux and MySQL, and prediction is evaluated based on Area Under Curve (AUC) performance measure. The key results of the proposed approach are as follows: It significantly gives better results of prediction performance for heterogeneous projects as compared with cross projects. Also, it demonstrates that feature selection with feature mapping has a significant effect on HFP models. Non-parametric statistical analyses, such as the Friedman and Nemenyi Post-hoc Tests, are applied, demonstrating that Logistic Regression performed significantly better than other supervised learning algorithms in HFP models.
{"title":"Heterogeneous Fault Prediction Using Feature Selection and Supervised Learning Algorithms","authors":"R. Arora, Arvinder Kaur","doi":"10.1142/s2196888822500142","DOIUrl":"https://doi.org/10.1142/s2196888822500142","url":null,"abstract":"Software Fault Prediction (SFP) is the most persuasive research area of software engineering. Software Fault Prediction which is carried out within the same software project is known as With-In Fault Prediction. However, local data repositories are not enough to build the model of With-in software Fault prediction. The idea of cross-project fault prediction (CPFP) has been suggested in recent years, which aims to construct a prediction model on one project, and use that model to predict the other project. However, CPFP requires that both the training and testing datasets use the same set of metrics. As a consequence, traditional CPFP approaches are challenging to implement through projects with diverse metric sets. The specific case of CPFP is Heterogeneous Fault Prediction (HFP), which allows the program to predict faults among projects with diverse metrics. The proposed framework aims to achieve an HFP model by implementing Feature Selection on both the source and target datasets to build an efficient prediction model using supervised machine learning techniques. Our approach is applied on two open-source projects, Linux and MySQL, and prediction is evaluated based on Area Under Curve (AUC) performance measure. The key results of the proposed approach are as follows: It significantly gives better results of prediction performance for heterogeneous projects as compared with cross projects. Also, it demonstrates that feature selection with feature mapping has a significant effect on HFP models. Non-parametric statistical analyses, such as the Friedman and Nemenyi Post-hoc Tests, are applied, demonstrating that Logistic Regression performed significantly better than other supervised learning algorithms in HFP models.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116940848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2022-01-20DOI: 10.1142/s2196888822500129
Ngoc Thang Nguyen, Van-Thanh Phan, Van Dat Nguyen, Thanh Ha Le, Thao Vy Pham
Forecasting the domestic coffee consumption demand is important for policy planning and making the right decisions. Thus, in this study, we try to find out the most suitable model among three proposed models (GM (1,1), DGM (1,1) and Grey Verhulst model (GVM)) for predicting the amount of domestic coffee consumption in Vietnam in the future. Yearly data of coffee consumption from 2010–2020 are used in this research. The experimental results indicated that the GM (1,1) is the most accurate model selected in this study with the lowest average value of [Formula: see text]%. So, the GM (1,1) model is strongly suggested in the analysis of coffee consumption demand in Vietnam. Finding the right tool will help managers make right decisions easily for sustainable development of the coffee industry in Vietnam in the future.
{"title":"Forecasting the Coffee Consumption Demand in Vietnam Based on Grey Forecasting Model","authors":"Ngoc Thang Nguyen, Van-Thanh Phan, Van Dat Nguyen, Thanh Ha Le, Thao Vy Pham","doi":"10.1142/s2196888822500129","DOIUrl":"https://doi.org/10.1142/s2196888822500129","url":null,"abstract":"Forecasting the domestic coffee consumption demand is important for policy planning and making the right decisions. Thus, in this study, we try to find out the most suitable model among three proposed models (GM (1,1), DGM (1,1) and Grey Verhulst model (GVM)) for predicting the amount of domestic coffee consumption in Vietnam in the future. Yearly data of coffee consumption from 2010–2020 are used in this research. The experimental results indicated that the GM (1,1) is the most accurate model selected in this study with the lowest average value of [Formula: see text]%. So, the GM (1,1) model is strongly suggested in the analysis of coffee consumption demand in Vietnam. Finding the right tool will help managers make right decisions easily for sustainable development of the coffee industry in Vietnam in the future.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"160 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123102677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-12-31DOI: 10.1142/s2196888822500130
Mithilesh Pandey, Sunita Jalal, Chetan S. Negi, D. Yadav
Due to the increasing number of Web Services with the same functionality, selecting a Web Service that best serves the needs of the Web Client has become a tremendously challenging task. Present approaches use non-functional parameters of the Web Services but they do not consider any preprocessing of the set of functionally Similar Web Services. The lack of preprocessing results in increased use of computational resources due to unnecessary processing of Web Services that have a very low to no chance of satisfying the consumer’s requirements. In this paper, we propose an Ensemble classification method for preprocessing and a Web Service Selection method based on the Quality of Service (QoS) parameters. Once the most eligible Web Services are enumerated through classification, they are ranked using the Technique of Order Preference by Similarity to Ideal Solution (TOPSIS) method with Analytic Hierarchy Process (AHP) used for weight calculation. A prototype of the method is developed, and experiments are conducted on a real-world Web Services dataset. Results demonstrate the feasibility of the proposed method.
由于具有相同功能的Web服务数量不断增加,选择最能满足Web客户机需求的Web服务已成为一项极具挑战性的任务。目前的方法使用Web服务的非功能参数,但它们不考虑对功能相似的Web服务集进行任何预处理。由于对Web服务进行不必要的处理,缺乏预处理会导致计算资源的使用增加,而这些Web服务几乎没有机会满足消费者的需求。本文提出了一种集成分类的预处理方法和一种基于服务质量(QoS)参数的Web服务选择方法。一旦通过分类列举出最符合条件的Web服务,就使用TOPSIS (Order Preference Technique of Similarity to Ideal Solution)方法对它们进行排序,并使用层次分析法(Analytic Hierarchy Process, AHP)进行权重计算。开发了该方法的原型,并在真实的Web Services数据集上进行了实验。结果证明了该方法的可行性。
{"title":"Using Ensemble and TOPSIS with AHP for Classification and Selection of Web Services","authors":"Mithilesh Pandey, Sunita Jalal, Chetan S. Negi, D. Yadav","doi":"10.1142/s2196888822500130","DOIUrl":"https://doi.org/10.1142/s2196888822500130","url":null,"abstract":"Due to the increasing number of Web Services with the same functionality, selecting a Web Service that best serves the needs of the Web Client has become a tremendously challenging task. Present approaches use non-functional parameters of the Web Services but they do not consider any preprocessing of the set of functionally Similar Web Services. The lack of preprocessing results in increased use of computational resources due to unnecessary processing of Web Services that have a very low to no chance of satisfying the consumer’s requirements. In this paper, we propose an Ensemble classification method for preprocessing and a Web Service Selection method based on the Quality of Service (QoS) parameters. Once the most eligible Web Services are enumerated through classification, they are ranked using the Technique of Order Preference by Similarity to Ideal Solution (TOPSIS) method with Analytic Hierarchy Process (AHP) used for weight calculation. A prototype of the method is developed, and experiments are conducted on a real-world Web Services dataset. Results demonstrate the feasibility of the proposed method.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125717113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-11-17DOI: 10.1142/s2196888822500105
Mohammad Alsawwaf, Z. Chaczko, Marek Kulbacki, Nikhil Sarathy
These days identification of a person is an integral part of many computer-based solutions. It is a key characteristic for access control, customized services, and a proof of identity. Over the last couple of decades, many new techniques were introduced for how to identify human faces. This approach investigates the human face identification based on frontal images by producing ratios from distances between the different features and their locations. Moreover, this extended version includes an investigation of identification based on side profile by extracting and diagnosing the feature sets with geometric ratio expressions which are calculated into feature vectors. The last stage involves using weighted means to calculate the resemblance. The approach considers an explainable Artificial Intelligence (XAI) approach. Findings, based on a small dataset, achieve that the used approach offers promising results. Further research could have a great influence on how faces and face-profiles can be identified. Performance of the proposed system is validated using metrics such as Precision, False Acceptance Rate, False Rejection Rate, and True Positive Rate. Multiple simulations indicate an Equal Error Rate of 0.89. This work is an extended version of the paper submitted in ACIIDS 2020.
{"title":"In Your Face: Person Identification Through Ratios and Distances Between Facial Features","authors":"Mohammad Alsawwaf, Z. Chaczko, Marek Kulbacki, Nikhil Sarathy","doi":"10.1142/s2196888822500105","DOIUrl":"https://doi.org/10.1142/s2196888822500105","url":null,"abstract":"These days identification of a person is an integral part of many computer-based solutions. It is a key characteristic for access control, customized services, and a proof of identity. Over the last couple of decades, many new techniques were introduced for how to identify human faces. This approach investigates the human face identification based on frontal images by producing ratios from distances between the different features and their locations. Moreover, this extended version includes an investigation of identification based on side profile by extracting and diagnosing the feature sets with geometric ratio expressions which are calculated into feature vectors. The last stage involves using weighted means to calculate the resemblance. The approach considers an explainable Artificial Intelligence (XAI) approach. Findings, based on a small dataset, achieve that the used approach offers promising results. Further research could have a great influence on how faces and face-profiles can be identified. Performance of the proposed system is validated using metrics such as Precision, False Acceptance Rate, False Rejection Rate, and True Positive Rate. Multiple simulations indicate an Equal Error Rate of 0.89. This work is an extended version of the paper submitted in ACIIDS 2020.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"156 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114408437","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-11-17DOI: 10.1142/s2196888822500075
Julia Garbaruk, D. Logofătu, C. Bǎdicǎ, Florin Leon
Whether for optimizing the speed of microprocessors or for sequence analysis in molecular biology — evolutionary algorithms are used in astoundingly many fields. Also, the art was influenced by evolutionary algorithms — with principles of natural evolution works of art that can be created or imitated, whereby initially generated art is put through an iterated process of selection and modification. This paper covers an application in which given images are emulated evolutionary using a finite number of semi-transparent overlapping polygons, which also became known under the name “Evolution of Mona Lisa”. In this context, different approaches to solve the problem are tested and presented here. In particular, we want to investigate whether Hill Climbing Algorithm in combination with Delaunay Triangulation and Canny Edge Detector that extracts the initial population directly from the original image performs better than the conventional Hill Climbing and Genetic Algorithm, where the initial population is generated randomly.
{"title":"Digital Image Evolution of Artwork Without Human Evaluation Using the Example of the Evolving Mona Lisa Problem","authors":"Julia Garbaruk, D. Logofătu, C. Bǎdicǎ, Florin Leon","doi":"10.1142/s2196888822500075","DOIUrl":"https://doi.org/10.1142/s2196888822500075","url":null,"abstract":"Whether for optimizing the speed of microprocessors or for sequence analysis in molecular biology — evolutionary algorithms are used in astoundingly many fields. Also, the art was influenced by evolutionary algorithms — with principles of natural evolution works of art that can be created or imitated, whereby initially generated art is put through an iterated process of selection and modification. This paper covers an application in which given images are emulated evolutionary using a finite number of semi-transparent overlapping polygons, which also became known under the name “Evolution of Mona Lisa”. In this context, different approaches to solve the problem are tested and presented here. In particular, we want to investigate whether Hill Climbing Algorithm in combination with Delaunay Triangulation and Canny Edge Detector that extracts the initial population directly from the original image performs better than the conventional Hill Climbing and Genetic Algorithm, where the initial population is generated randomly.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124954374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-10-23DOI: 10.1142/s2196888822500117
C. C. Benson, V. Lajish, K. Rajamani
Fully automatic brain image classification of MR brain images is of great importance for research and clinical studies, since the precise detection may lead to a better treatment. In this work, an efficient method based on Multiple-Instance Learning (MIL) is proposed for the automatic classification of low-grade and high-grade MR brain tumor images. The main advantage of MIL-based approach over other classification methods is that MIL considers an image as a group of instances rather than a single instance, thus facilitating an effective learning process. The mi-Graph-based MIL approach is proposed for this classification. Two different implementations of MIL-based classification, viz. Patch-based MIL (PBMIL) and Superpixel-based MIL (SPBMIL), are made in this study. The combined feature set of LBP, SIFT and FD is used for the classification. The accuracies of low-grade–high-grade tumor image classification algorithm using SPBMIL method performed on [Formula: see text], [Formula: see text] and FLAIR images read 99.2765%, 99.4195% and 99.2265%, respectively. The error rate of the proposed classification system was noted to be insignificant and hence this automated classification system could be used for the classification of images with different pathological conditions, types and disease statuses.
{"title":"Classification of Low-Grade and High-Grade Glioma from MR Brain Images Using Multiple-Instance Learning with Combined Feature Set","authors":"C. C. Benson, V. Lajish, K. Rajamani","doi":"10.1142/s2196888822500117","DOIUrl":"https://doi.org/10.1142/s2196888822500117","url":null,"abstract":"Fully automatic brain image classification of MR brain images is of great importance for research and clinical studies, since the precise detection may lead to a better treatment. In this work, an efficient method based on Multiple-Instance Learning (MIL) is proposed for the automatic classification of low-grade and high-grade MR brain tumor images. The main advantage of MIL-based approach over other classification methods is that MIL considers an image as a group of instances rather than a single instance, thus facilitating an effective learning process. The mi-Graph-based MIL approach is proposed for this classification. Two different implementations of MIL-based classification, viz. Patch-based MIL (PBMIL) and Superpixel-based MIL (SPBMIL), are made in this study. The combined feature set of LBP, SIFT and FD is used for the classification. The accuracies of low-grade–high-grade tumor image classification algorithm using SPBMIL method performed on [Formula: see text], [Formula: see text] and FLAIR images read 99.2765%, 99.4195% and 99.2265%, respectively. The error rate of the proposed classification system was noted to be insignificant and hence this automated classification system could be used for the classification of images with different pathological conditions, types and disease statuses.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131119603","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-10-06DOI: 10.1142/s219688882250004x
A. Babhulgaonkar, S. Sonavane
Hindi is the national language of India. However, most of the Government records, resolutions, news, etc. are documented in English which remote villagers may not understand. This fact motivates to develop an automatic language translation system from English to Hindi. Machine translation is the process of translating a text in one natural language into another natural language using computer system. Grammatical structure of Hindi language is very much complex than English language. The structural difference between English and Hindi language makes it difficult to achieve good quality translation results. In this paper, phrase-based statistical machine translation approach (PBSMT) is used for translation. Translation, reordering and language model are main working components of a PBSMT system. This paper evaluates the impact of various combinations of these PBSMT system parameters on automated English to Hindi language translation quality. Freely available n-gram-based BLEU metric and TER metric are used for evaluating the results.
{"title":"Empirical Analysis of Phrase-Based Statistical Machine Translation System for English to Hindi Language","authors":"A. Babhulgaonkar, S. Sonavane","doi":"10.1142/s219688882250004x","DOIUrl":"https://doi.org/10.1142/s219688882250004x","url":null,"abstract":"Hindi is the national language of India. However, most of the Government records, resolutions, news, etc. are documented in English which remote villagers may not understand. This fact motivates to develop an automatic language translation system from English to Hindi. Machine translation is the process of translating a text in one natural language into another natural language using computer system. Grammatical structure of Hindi language is very much complex than English language. The structural difference between English and Hindi language makes it difficult to achieve good quality translation results. In this paper, phrase-based statistical machine translation approach (PBSMT) is used for translation. Translation, reordering and language model are main working components of a PBSMT system. This paper evaluates the impact of various combinations of these PBSMT system parameters on automated English to Hindi language translation quality. Freely available n-gram-based BLEU metric and TER metric are used for evaluating the results.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-10-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125984257","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2021-09-18DOI: 10.1142/s2196888822500087
Md. Tarek Habib, Md. Jueal Mia, Mohammad Shorif Uddin, F. Ahmed
Bangladesh, being a densely populated country, hinges on agriculture for the security of finance and food to a large extent. Hence, both the fruits’ quantity and quality turn out to be very important, which can be degraded due to the attacks of various diseases. Automated fruit disease recognition can help fruit farmers, especially remote farmers, for whom adequate cultivation support is required. Two daunting problems, namely disease detection, and disease classification are raised by automated fruit disease recognition. In this research, we conduct an intense investigation of the applicability of automated recognition of the diseases of three available Bangladeshi local fruits, viz. guava, jackfruit, and papaya. After exerting four notable segmentation algorithms, [Formula: see text]-means clustering segmentation algorithm is selected to segregate the disease-contaminated parts from a fruit image. Then some discriminatory features are extracted from these disease-contaminated parts. Nine noteworthy classification algorithms are applied for disease classification to thoroughly get the measure of their merits. It is observed that random forest outperforms the eight other classifiers by disclosing an accuracy of 96.8% and 89.59% for guava and jackfruit, respectively, whereas support vector machine attains an accuracy of 94.9% for papaya, which can be claimed good as well as attractive for forthcoming research.
{"title":"An Explorative Analysis on the Machine-Vision-Based Disease Recognition of Three Available Fruits of Bangladesh","authors":"Md. Tarek Habib, Md. Jueal Mia, Mohammad Shorif Uddin, F. Ahmed","doi":"10.1142/s2196888822500087","DOIUrl":"https://doi.org/10.1142/s2196888822500087","url":null,"abstract":"Bangladesh, being a densely populated country, hinges on agriculture for the security of finance and food to a large extent. Hence, both the fruits’ quantity and quality turn out to be very important, which can be degraded due to the attacks of various diseases. Automated fruit disease recognition can help fruit farmers, especially remote farmers, for whom adequate cultivation support is required. Two daunting problems, namely disease detection, and disease classification are raised by automated fruit disease recognition. In this research, we conduct an intense investigation of the applicability of automated recognition of the diseases of three available Bangladeshi local fruits, viz. guava, jackfruit, and papaya. After exerting four notable segmentation algorithms, [Formula: see text]-means clustering segmentation algorithm is selected to segregate the disease-contaminated parts from a fruit image. Then some discriminatory features are extracted from these disease-contaminated parts. Nine noteworthy classification algorithms are applied for disease classification to thoroughly get the measure of their merits. It is observed that random forest outperforms the eight other classifiers by disclosing an accuracy of 96.8% and 89.59% for guava and jackfruit, respectively, whereas support vector machine attains an accuracy of 94.9% for papaya, which can be claimed good as well as attractive for forthcoming research.","PeriodicalId":256649,"journal":{"name":"Vietnam. J. Comput. Sci.","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2021-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121557410","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}