The growth in the number of people suffering from eye diseases determines the relevance of research in the field of diagnosing retinal pathologies. Artificial intelligence models and algorithms based on measurements obtained via electrophysiological methods can significantly improve and speed up the analysis of results and diagnostics. We propose an approach to designing an artificial intelligent diagnosis system (AI diagnosis system) which includes an electrophysiological complex to collect objective information and an intelligent decision support system to justify the diagnosis. The task of diagnosing retinal diseases based on a set of heterogeneous data is considered as a multi-class classification on unbalanced data. The decision support system includes two classifiers—one classifier is based on a fuzzy model and a fuzzy rule base (RB-classifier) and one uses the stochastic gradient boosting algorithm (SGB-classifier). The efficiency of algorithms in a multi-class classification on unbalanced data is assessed based on two indicators—MAUC (multi-class area under curve) and MMCC (multi-class Matthews correlation coefficient). Combining two algorithms in a decision support system provides more accurate and reliable pathology identification. The accuracy of diagnostics using the proposed AI diagnosis system is 5–8% higher than the accuracy of a system using only diagnostics based on electrophysical indicators. The AI diagnosis system differs from other systems of this class in that it is based on the processing of objective electrophysiological data and socio-demographic data about patients, as well as subjective information from the anamnesis, which ensures increased efficiency of medical decision-making. The system is tested using actual data about retinal diseases from the Russian Institute of Eye Diseases and its high efficiency is proven. Simulation experiments conducted in various scenario conditions with different combinations of factors ensured the identification of the main determinants (markers) for each diagnosis of retinal pathology.
{"title":"Artificial Intelligence-Based System for Retinal Disease Diagnosis","authors":"E. V. Orlova","doi":"10.3390/a17070315","DOIUrl":"https://doi.org/10.3390/a17070315","url":null,"abstract":"The growth in the number of people suffering from eye diseases determines the relevance of research in the field of diagnosing retinal pathologies. Artificial intelligence models and algorithms based on measurements obtained via electrophysiological methods can significantly improve and speed up the analysis of results and diagnostics. We propose an approach to designing an artificial intelligent diagnosis system (AI diagnosis system) which includes an electrophysiological complex to collect objective information and an intelligent decision support system to justify the diagnosis. The task of diagnosing retinal diseases based on a set of heterogeneous data is considered as a multi-class classification on unbalanced data. The decision support system includes two classifiers—one classifier is based on a fuzzy model and a fuzzy rule base (RB-classifier) and one uses the stochastic gradient boosting algorithm (SGB-classifier). The efficiency of algorithms in a multi-class classification on unbalanced data is assessed based on two indicators—MAUC (multi-class area under curve) and MMCC (multi-class Matthews correlation coefficient). Combining two algorithms in a decision support system provides more accurate and reliable pathology identification. The accuracy of diagnostics using the proposed AI diagnosis system is 5–8% higher than the accuracy of a system using only diagnostics based on electrophysical indicators. The AI diagnosis system differs from other systems of this class in that it is based on the processing of objective electrophysiological data and socio-demographic data about patients, as well as subjective information from the anamnesis, which ensures increased efficiency of medical decision-making. The system is tested using actual data about retinal diseases from the Russian Institute of Eye Diseases and its high efficiency is proven. Simulation experiments conducted in various scenario conditions with different combinations of factors ensured the identification of the main determinants (markers) for each diagnosis of retinal pathology.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":" 12","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141824561","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Itzel M. Abundez, Roberto Alejo, Francisco Primero Primero, E. Granda-Gutiérrez, O. Portillo-Rodríguez, Juan Alberto Antonio Velázquez
Public authorities and private companies have used video cameras as part of surveillance systems, and one of their objectives is the rapid detection of physically violent actions. This task is usually performed by human visual inspection, which is labor-intensive. For this reason, different deep learning models have been implemented to remove the human eye from this task, yielding positive results. One of the main problems in detecting physical violence in videos is the variety of scenarios that can exist, which leads to different models being trained on datasets, leading them to detect physical violence in only one or a few types of videos. In this work, we present an approach for physical violence detection on images obtained from video based on threshold active learning, that increases the classifier’s robustness in environments where it was not trained. The proposed approach consists of two stages: In the first stage, pre-trained neural network models are trained on initial datasets, and we use a threshold (μ) to identify those images that the classifier considers ambiguous or hard to classify. Then, they are included in the training dataset, and the model is retrained to improve its classification performance. In the second stage, we test the model with video images from other environments, and we again employ (μ) to detect ambiguous images that a human expert analyzes to determine the real class or delete the ambiguity on them. After that, the ambiguous images are added to the original training set and the classifier is retrained; this process is repeated while ambiguous images exist. The model is a hybrid neural network that uses transfer learning and a threshold μ to detect physical violence on images obtained from video files successfully. In this active learning process, the classifier can detect physical violence in different environments, where the main contribution is the method used to obtain a threshold μ (which is based on the neural network output) that allows human experts to contribute to the classification process to obtain more robust neural networks and high-quality datasets. The experimental results show the proposed approach’s effectiveness in detecting physical violence, where it is trained using an initial dataset, and new images are added to improve its robustness in diverse environments.
{"title":"Threshold Active Learning Approach for Physical Violence Detection on Images Obtained from Video (Frame-Level) Using Pre-Trained Deep Learning Neural Network Models","authors":"Itzel M. Abundez, Roberto Alejo, Francisco Primero Primero, E. Granda-Gutiérrez, O. Portillo-Rodríguez, Juan Alberto Antonio Velázquez","doi":"10.3390/a17070316","DOIUrl":"https://doi.org/10.3390/a17070316","url":null,"abstract":"Public authorities and private companies have used video cameras as part of surveillance systems, and one of their objectives is the rapid detection of physically violent actions. This task is usually performed by human visual inspection, which is labor-intensive. For this reason, different deep learning models have been implemented to remove the human eye from this task, yielding positive results. One of the main problems in detecting physical violence in videos is the variety of scenarios that can exist, which leads to different models being trained on datasets, leading them to detect physical violence in only one or a few types of videos. In this work, we present an approach for physical violence detection on images obtained from video based on threshold active learning, that increases the classifier’s robustness in environments where it was not trained. The proposed approach consists of two stages: In the first stage, pre-trained neural network models are trained on initial datasets, and we use a threshold (μ) to identify those images that the classifier considers ambiguous or hard to classify. Then, they are included in the training dataset, and the model is retrained to improve its classification performance. In the second stage, we test the model with video images from other environments, and we again employ (μ) to detect ambiguous images that a human expert analyzes to determine the real class or delete the ambiguity on them. After that, the ambiguous images are added to the original training set and the classifier is retrained; this process is repeated while ambiguous images exist. The model is a hybrid neural network that uses transfer learning and a threshold μ to detect physical violence on images obtained from video files successfully. In this active learning process, the classifier can detect physical violence in different environments, where the main contribution is the method used to obtain a threshold μ (which is based on the neural network output) that allows human experts to contribute to the classification process to obtain more robust neural networks and high-quality datasets. The experimental results show the proposed approach’s effectiveness in detecting physical violence, where it is trained using an initial dataset, and new images are added to improve its robustness in diverse environments.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":" 8","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-07-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141827195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
With the rapid development of computer technology, communication technology, and control technology, cyber-physical systems (CPSs) have been widely used and developed. However, there are massive information interactions in CPSs, which lead to an increase in the amount of data transmitted over the network. The data communication, once attacked by the network, will seriously affect the security and stability of the system. In this paper, for the data tampering attack existing in the linear system with multiple binary observations, in the case where the estimation algorithm of the defender is unknown, the optimization index is constructed based on information entropy from the attacker’s point of view, and the problem is modeled. For the problem of the multi-parameter optimization with energy constraints, this paper uses particle swarm optimization (PSO) to obtain the optimal data tampering attack solution set, and gives the estimation method of unknown parameters in the case of unknown parameters. To implement the real-time improvement of online implementation, the BP neural network is designed. Finally, the validity of the conclusions is verified through numerical simulation. This means that the attacker can construct effective metrics based on information entropy without the knowledge of the defense’s discrimination algorithm. In addition, the optimal attack strategy implementation based on PSO and BP is also effective.
随着计算机技术、通信技术和控制技术的飞速发展,网络物理系统(CPS)得到了广泛的应用和发展。然而,CPS 中存在大量的信息交互,导致网络传输的数据量不断增加。数据通信一旦受到网络攻击,将严重影响系统的安全性和稳定性。本文针对具有多个二进制观测值的线性系统中存在的数据篡改攻击,在防御方估计算法未知的情况下,从攻击方的角度出发,构建了基于信息熵的优化指标,并对问题进行了建模。针对有能量约束的多参数优化问题,本文利用粒子群优化(PSO)获得最优数据篡改攻击解集,并给出了未知参数情况下的未知参数估计方法。为了实现在线执行的实时改进,设计了 BP 神经网络。最后,通过数值模拟验证了结论的正确性。这意味着攻击方可以在不知道防御方判别算法的情况下,根据信息熵构建有效的度量。此外,基于 PSO 和 BP 的最优攻击策略实施也是有效的。
{"title":"Linear System Identification-Oriented Optimal Tampering Attack Strategy and Implementation Based on Information Entropy with Multiple Binary Observations","authors":"Zhongwei Bai, Peng Yu, Yan Liu, Jin Guo","doi":"10.3390/a17060239","DOIUrl":"https://doi.org/10.3390/a17060239","url":null,"abstract":"With the rapid development of computer technology, communication technology, and control technology, cyber-physical systems (CPSs) have been widely used and developed. However, there are massive information interactions in CPSs, which lead to an increase in the amount of data transmitted over the network. The data communication, once attacked by the network, will seriously affect the security and stability of the system. In this paper, for the data tampering attack existing in the linear system with multiple binary observations, in the case where the estimation algorithm of the defender is unknown, the optimization index is constructed based on information entropy from the attacker’s point of view, and the problem is modeled. For the problem of the multi-parameter optimization with energy constraints, this paper uses particle swarm optimization (PSO) to obtain the optimal data tampering attack solution set, and gives the estimation method of unknown parameters in the case of unknown parameters. To implement the real-time improvement of online implementation, the BP neural network is designed. Finally, the validity of the conclusions is verified through numerical simulation. This means that the attacker can construct effective metrics based on information entropy without the knowledge of the defense’s discrimination algorithm. In addition, the optimal attack strategy implementation based on PSO and BP is also effective.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":"31 49","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141270475","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Jirayu Samkunta, P. Ketthong, N. T. Mai, M.A.S. Kamal, I. Murakami, Kou Yamada
The kinematics of the human hand exhibit complex and diverse characteristics unique to each individual. Various techniques such as vision-based, ultrasonic-based, and data-glove-based approaches have been employed to analyze human hand movements. However, a critical challenge remains in efficiently analyzing and classifying hand grasp types based on time-series kinematic data. In this paper, we propose a novel sparse coding feature extraction technique based on dictionary learning to address this challenge. Our method enhances model accuracy, reduces training time, and minimizes overfitting risk. We benchmarked our approach against principal component analysis (PCA) and sparse coding based on a Gaussian random dictionary. Our results demonstrate a significant improvement in classification accuracy: achieving 81.78% with our method compared to 31.43% for PCA and 77.27% for the Gaussian random dictionary. Furthermore, our technique outperforms in terms of macro-average F1-score and average area under the curve (AUC) while also significantly reducing the number of features required.
{"title":"Feature Extraction Based on Sparse Coding Approach for Hand Grasp Type Classification","authors":"Jirayu Samkunta, P. Ketthong, N. T. Mai, M.A.S. Kamal, I. Murakami, Kou Yamada","doi":"10.3390/a17060240","DOIUrl":"https://doi.org/10.3390/a17060240","url":null,"abstract":"The kinematics of the human hand exhibit complex and diverse characteristics unique to each individual. Various techniques such as vision-based, ultrasonic-based, and data-glove-based approaches have been employed to analyze human hand movements. However, a critical challenge remains in efficiently analyzing and classifying hand grasp types based on time-series kinematic data. In this paper, we propose a novel sparse coding feature extraction technique based on dictionary learning to address this challenge. Our method enhances model accuracy, reduces training time, and minimizes overfitting risk. We benchmarked our approach against principal component analysis (PCA) and sparse coding based on a Gaussian random dictionary. Our results demonstrate a significant improvement in classification accuracy: achieving 81.78% with our method compared to 31.43% for PCA and 77.27% for the Gaussian random dictionary. Furthermore, our technique outperforms in terms of macro-average F1-score and average area under the curve (AUC) while also significantly reducing the number of features required.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":"49 50","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141269734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We study an online distribution problem in which a producer has to send a load from an origin to a destination. At each time period before the deadline, they ask for transportation price quotes and have to decide to either accept or not accept the minimum offered price. If this price is not accepted, they have to pay a penalty cost, which may be the cost to ask for new quotes, the penalty cost for a late delivery, or the inventory cost to store the load for a certain duration. The aim is to minimize the sum of the transportation and the penalty costs. This problem has interesting real-world applications, given that transportation quotes can be obtained from professional websites nowadays. We show that the classical online algorithm used to solve the well-known Secretary problem is not able to provide, on average, effective solutions to our problem, given the trade-off between the transportation and the penalty costs. Therefore, we design two classes of online algorithms. The first class is based on a given time of acceptance, while the second is based on a given threshold price. We formally prove the competitive ratio of each algorithm, i.e., the worst-case performance of the online algorithm with respect to the optimal solution of the offline problem, in which all transportation prices are known at the beginning, rather than being revealed over time. The computational results show the algorithms’ performance on average and in the worst-case scenario when the transportation prices are generated on the basis of given probability distributions.
{"title":"Competitive Analysis of Algorithms for an Online Distribution Problem","authors":"Alessandro Barba, L. Bertazzi, Bruce L. Golden","doi":"10.3390/a17060237","DOIUrl":"https://doi.org/10.3390/a17060237","url":null,"abstract":"We study an online distribution problem in which a producer has to send a load from an origin to a destination. At each time period before the deadline, they ask for transportation price quotes and have to decide to either accept or not accept the minimum offered price. If this price is not accepted, they have to pay a penalty cost, which may be the cost to ask for new quotes, the penalty cost for a late delivery, or the inventory cost to store the load for a certain duration. The aim is to minimize the sum of the transportation and the penalty costs. This problem has interesting real-world applications, given that transportation quotes can be obtained from professional websites nowadays. We show that the classical online algorithm used to solve the well-known Secretary problem is not able to provide, on average, effective solutions to our problem, given the trade-off between the transportation and the penalty costs. Therefore, we design two classes of online algorithms. The first class is based on a given time of acceptance, while the second is based on a given threshold price. We formally prove the competitive ratio of each algorithm, i.e., the worst-case performance of the online algorithm with respect to the optimal solution of the offline problem, in which all transportation prices are known at the beginning, rather than being revealed over time. The computational results show the algorithms’ performance on average and in the worst-case scenario when the transportation prices are generated on the basis of given probability distributions.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":"116 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141271908","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Facial emotion recognition (FER) is crucial across psychology, neuroscience, computer vision, and machine learning due to the diversified and subjective nature of emotions, varying considerably across individuals, cultures, and contexts. This study explored FER through convolutional neural networks (CNNs) and Histogram Equalization techniques. It investigated the impact of histogram equalization, data augmentation, and various model optimization strategies on FER accuracy across different datasets like KDEF, CK+, and FER2013. Using pre-trained VGG architectures, such as VGG19 and VGG16, this study also examined the effectiveness of fine-tuning hyperparameters and implementing different learning rate schedulers. The evaluation encompassed diverse metrics including accuracy, Area Under the Receiver Operating Characteristic Curve (AUC-ROC), Area Under the Precision–Recall Curve (AUC-PRC), and Weighted F1 score. Notably, the fine-tuned VGG architecture demonstrated a state-of-the-art performance compared to conventional transfer learning models and achieved 100%, 95.92%, and 69.65% on the CK+, KDEF, and FER2013 datasets, respectively.
面部情绪识别(FER)在心理学、神经科学、计算机视觉和机器学习领域都至关重要,因为情绪具有多样化和主观性的特点,不同的个体、文化和环境会有很大的差异。本研究通过卷积神经网络(CNN)和直方图均衡化技术对 FER 进行了探索。它研究了直方图均衡化、数据增强和各种模型优化策略对 KDEF、CK+ 和 FER2013 等不同数据集的 FER 准确率的影响。这项研究还使用预先训练好的 VGG 架构(如 VGG19 和 VGG16),检验了微调超参数和实施不同学习率调度器的效果。评估涵盖了多种指标,包括准确率、接收者操作特征曲线下面积(AUC-ROC)、精度-召回曲线下面积(AUC-PRC)和加权 F1 分数。值得注意的是,与传统的迁移学习模型相比,经过微调的 VGG 架构表现出了最先进的性能,在 CK+、KDEF 和 FER2013 数据集上的准确率分别达到了 100%、95.92% 和 69.65%。
{"title":"Simple Histogram Equalization Technique Improves Performance of VGG Models on Facial Emotion Recognition Datasets","authors":"Jaher Hassan Chowdhury, Qian Liu, S. Ramanna","doi":"10.3390/a17060238","DOIUrl":"https://doi.org/10.3390/a17060238","url":null,"abstract":"Facial emotion recognition (FER) is crucial across psychology, neuroscience, computer vision, and machine learning due to the diversified and subjective nature of emotions, varying considerably across individuals, cultures, and contexts. This study explored FER through convolutional neural networks (CNNs) and Histogram Equalization techniques. It investigated the impact of histogram equalization, data augmentation, and various model optimization strategies on FER accuracy across different datasets like KDEF, CK+, and FER2013. Using pre-trained VGG architectures, such as VGG19 and VGG16, this study also examined the effectiveness of fine-tuning hyperparameters and implementing different learning rate schedulers. The evaluation encompassed diverse metrics including accuracy, Area Under the Receiver Operating Characteristic Curve (AUC-ROC), Area Under the Precision–Recall Curve (AUC-PRC), and Weighted F1 score. Notably, the fine-tuned VGG architecture demonstrated a state-of-the-art performance compared to conventional transfer learning models and achieved 100%, 95.92%, and 69.65% on the CK+, KDEF, and FER2013 datasets, respectively.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":"60 20","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141269498","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The adequacy and efficacy of simple and hybrid machine learning and Computational Intelligence algorithms were evaluated for the classification of potential prostate cancer patients in two distinct categories, the high- and the low-risk group for PCa. The evaluation is based on randomly generated surrogate data for the biomarker PSA, considering that reported epidemiological data indicated that PSA values follow a lognormal distribution. In addition, four more biomarkers were considered, namely, PSAD (PSA density), PSAV (PSA velocity), PSA ratio, and Digital Rectal Exam evaluation (DRE), as well as patient age. Seven simple classification algorithms, namely, Decision Trees, Random Forests, Support Vector Machines, K-Nearest Neighbors, Logistic Regression, Naïve Bayes, and Artificial Neural Networks, were evaluated in terms of classification accuracy. In addition, three hybrid algorithms were developed and introduced in the present work, where Genetic Algorithms were utilized as a metaheuristic searching technique in order to optimize the training set, in terms of minimizing its size, to give optimal classification accuracy for the simple algorithms including K-Nearest Neighbors, a K-means clustering algorithm, and a genetic clustering algorithm. Results indicated that prostate cancer cases can be classified with high accuracy, even by the use of small training sets, with sizes that could be even smaller than 30% of the dataset. Numerous computer experiments indicated that the proposed training set minimization does not cause overfitting of the hybrid algorithms. Finally, an easy-to-use Graphical User Interface (GUI) was implemented, incorporating all the evaluated algorithms and the decision-making procedure.
{"title":"Hybrid Machine Learning Algorithms to Evaluate Prostate Cancer","authors":"Dimitrios Morakis, Adam Adamopoulos","doi":"10.3390/a17060236","DOIUrl":"https://doi.org/10.3390/a17060236","url":null,"abstract":"The adequacy and efficacy of simple and hybrid machine learning and Computational Intelligence algorithms were evaluated for the classification of potential prostate cancer patients in two distinct categories, the high- and the low-risk group for PCa. The evaluation is based on randomly generated surrogate data for the biomarker PSA, considering that reported epidemiological data indicated that PSA values follow a lognormal distribution. In addition, four more biomarkers were considered, namely, PSAD (PSA density), PSAV (PSA velocity), PSA ratio, and Digital Rectal Exam evaluation (DRE), as well as patient age. Seven simple classification algorithms, namely, Decision Trees, Random Forests, Support Vector Machines, K-Nearest Neighbors, Logistic Regression, Naïve Bayes, and Artificial Neural Networks, were evaluated in terms of classification accuracy. In addition, three hybrid algorithms were developed and introduced in the present work, where Genetic Algorithms were utilized as a metaheuristic searching technique in order to optimize the training set, in terms of minimizing its size, to give optimal classification accuracy for the simple algorithms including K-Nearest Neighbors, a K-means clustering algorithm, and a genetic clustering algorithm. Results indicated that prostate cancer cases can be classified with high accuracy, even by the use of small training sets, with sizes that could be even smaller than 30% of the dataset. Numerous computer experiments indicated that the proposed training set minimization does not cause overfitting of the hybrid algorithms. Finally, an easy-to-use Graphical User Interface (GUI) was implemented, incorporating all the evaluated algorithms and the decision-making procedure.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":"25 23","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-06-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141272773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
This paper presents a novel approach to automated music mixing, focusing on the optimization of loudness control in multi-track recordings. By taking into consideration the complexity and artistic nature of traditional mixing processes, we introduce a personalized multi-track leveling method using two types of approaches: a customized genetic algorithm and a neural network-based method. Our method tackles common challenges encountered by audio professionals during prolonged mixing sessions, where consistency can decrease as a result of fatigue. Our algorithm serves as a ‘virtual assistant’ to consistently uphold the initial mixing objectives, hence assuring consistent quality throughout the process. In addition, our system automates the repetitive elements of the mixing process, resulting in a substantial reduction in production time. This enables engineers to dedicate their attention to more innovative and intricate jobs. Our experimental framework involves 20 diverse songs and 10 sound engineers possessing a wide range of expertise, offering a useful perspective on the adaptability and effectiveness of our method in real-world scenarios. The results demonstrate the capacity of the algorithms to mimic decision-making, achieving an optimal balance in the mix that resonates with the emotional and technical aspects of music production.
{"title":"Automated Personalized Loudness Control for Multi-Track Recordings","authors":"Bogdan Moroșanu, Marian Negru, C. Paleologu","doi":"10.3390/a17060228","DOIUrl":"https://doi.org/10.3390/a17060228","url":null,"abstract":"This paper presents a novel approach to automated music mixing, focusing on the optimization of loudness control in multi-track recordings. By taking into consideration the complexity and artistic nature of traditional mixing processes, we introduce a personalized multi-track leveling method using two types of approaches: a customized genetic algorithm and a neural network-based method. Our method tackles common challenges encountered by audio professionals during prolonged mixing sessions, where consistency can decrease as a result of fatigue. Our algorithm serves as a ‘virtual assistant’ to consistently uphold the initial mixing objectives, hence assuring consistent quality throughout the process. In addition, our system automates the repetitive elements of the mixing process, resulting in a substantial reduction in production time. This enables engineers to dedicate their attention to more innovative and intricate jobs. Our experimental framework involves 20 diverse songs and 10 sound engineers possessing a wide range of expertise, offering a useful perspective on the adaptability and effectiveness of our method in real-world scenarios. The results demonstrate the capacity of the algorithms to mimic decision-making, achieving an optimal balance in the mix that resonates with the emotional and technical aspects of music production.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":"51 2","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141102211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Khwansiri Ninpan, Shuzhang Huang, Francesco Vitillo, Mohamad Ali Assaad, Lies Benmiloud Bechet, Robert Plana
This paper proposes a heuristic approach for managing complex construction projects. The tool incorporates Primavera P6 EPPM and Synchro 4D, enabling proactive clash detection and resolution of spatial conflicts during concurrent tasks. Additionally, it performs resource verification for sufficient allocation before task initiation. This integrated approach facilitates the generation of conflict-free and feasible construction schedules. By adhering to project constraints and seamlessly integrating with existing industry tools, the proposed solution offers a comprehensive and robust approach to construction project management. This constitutes, to our knowledge, the first dynamic digital twin for the delivery of a complex project.
{"title":"Mitigating Co-Activity Conflicts and Resource Overallocation in Construction Projects: A Modular Heuristic Scheduling Approach with Primavera P6 EPPM Integration","authors":"Khwansiri Ninpan, Shuzhang Huang, Francesco Vitillo, Mohamad Ali Assaad, Lies Benmiloud Bechet, Robert Plana","doi":"10.3390/a17060230","DOIUrl":"https://doi.org/10.3390/a17060230","url":null,"abstract":"This paper proposes a heuristic approach for managing complex construction projects. The tool incorporates Primavera P6 EPPM and Synchro 4D, enabling proactive clash detection and resolution of spatial conflicts during concurrent tasks. Additionally, it performs resource verification for sufficient allocation before task initiation. This integrated approach facilitates the generation of conflict-free and feasible construction schedules. By adhering to project constraints and seamlessly integrating with existing industry tools, the proposed solution offers a comprehensive and robust approach to construction project management. This constitutes, to our knowledge, the first dynamic digital twin for the delivery of a complex project.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":"66 23","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141101663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Neeraj Anand Sharma, Rishal Ravikesh Chand, Zain Buksh, A. B. M. S. Ali, Ambreen Hanif, A. Beheshti
This study delves into the realm of Explainable Artificial Intelligence (XAI) frameworks, aiming to empower researchers and practitioners with a deeper understanding of these tools. We establish a comprehensive knowledge base by classifying and analyzing prominent XAI solutions based on key attributes like explanation type, model dependence, and use cases. This resource equips users to navigate the diverse XAI landscape and select the most suitable framework for their specific needs. Furthermore, the study proposes a novel framework called XAIE (eXplainable AI Evaluator) for informed decision-making in XAI adoption. This framework empowers users to assess different XAI options based on their application context objectively. This will lead to more responsible AI development by fostering transparency and trust. Finally, the research identifies the limitations and challenges associated with the existing XAI frameworks, paving the way for future advancements. By highlighting these areas, the study guides researchers and developers in enhancing the capabilities of Explainable AI.
{"title":"Explainable AI Frameworks: Navigating the Present Challenges and Unveiling Innovative Applications","authors":"Neeraj Anand Sharma, Rishal Ravikesh Chand, Zain Buksh, A. B. M. S. Ali, Ambreen Hanif, A. Beheshti","doi":"10.3390/a17060227","DOIUrl":"https://doi.org/10.3390/a17060227","url":null,"abstract":"This study delves into the realm of Explainable Artificial Intelligence (XAI) frameworks, aiming to empower researchers and practitioners with a deeper understanding of these tools. We establish a comprehensive knowledge base by classifying and analyzing prominent XAI solutions based on key attributes like explanation type, model dependence, and use cases. This resource equips users to navigate the diverse XAI landscape and select the most suitable framework for their specific needs. Furthermore, the study proposes a novel framework called XAIE (eXplainable AI Evaluator) for informed decision-making in XAI adoption. This framework empowers users to assess different XAI options based on their application context objectively. This will lead to more responsible AI development by fostering transparency and trust. Finally, the research identifies the limitations and challenges associated with the existing XAI frameworks, paving the way for future advancements. By highlighting these areas, the study guides researchers and developers in enhancing the capabilities of Explainable AI.","PeriodicalId":502609,"journal":{"name":"Algorithms","volume":"10 5","pages":""},"PeriodicalIF":0.0,"publicationDate":"2024-05-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141099486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}