首页 > 最新文献

Advances in Data Science and Adaptive Analysis最新文献

英文 中文
A New Class of Cure Rate Survival Models: Properties, Inference and Applications 一类新的治愈率生存模型:性质、推论及应用
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2021-03-27 DOI: 10.1142/S2424922X21500017
Gladys D. C. Barriga, A. K. Suzuki, V. Cancho, F. Louzada
We propose a new class of survival models for time-to-event data with a cure fraction. This new model is an extension of the promotion time cure rate model. Furthermore, we extend the model to the regression model to evaluate the effect of covariates in the cure fraction. An expectation-maximization algorithm is adopted for estimating the model parameters. A simulation study is conducted in order to assess the proposed model and the computation algorithm. The methodology is illustrated using a real Brazilian bank personal loan portfolio data set.
我们提出了一类新的生存模型的时间到事件的数据与治愈分数。这个新模型是对推广时间治愈率模型的扩展。进一步,我们将模型推广到回归模型,以评估协变量对治愈分数的影响。采用期望最大化算法对模型参数进行估计。为了验证所提出的模型和计算算法,进行了仿真研究。该方法使用一个真实的巴西银行个人贷款组合数据集来说明。
{"title":"A New Class of Cure Rate Survival Models: Properties, Inference and Applications","authors":"Gladys D. C. Barriga, A. K. Suzuki, V. Cancho, F. Louzada","doi":"10.1142/S2424922X21500017","DOIUrl":"https://doi.org/10.1142/S2424922X21500017","url":null,"abstract":"We propose a new class of survival models for time-to-event data with a cure fraction. This new model is an extension of the promotion time cure rate model. Furthermore, we extend the model to the regression model to evaluate the effect of covariates in the cure fraction. An expectation-maximization algorithm is adopted for estimating the model parameters. A simulation study is conducted in order to assess the proposed model and the computation algorithm. The methodology is illustrated using a real Brazilian bank personal loan portfolio data set.","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"35 1","pages":"2150001:1-2150001:15"},"PeriodicalIF":0.6,"publicationDate":"2021-03-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75969945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Empirical Spatial Density-Based Emergency Medical Service Demand Forecast for Ambulance Allocation 基于空间密度的救护车配置急诊医疗服务需求预测
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2021-02-17 DOI: 10.1142/S2424922X21500030
Y. Tsai, Hwei-Jen Lin, Pei Chi, Kelvin W. Lee
In a previous study, we solved the two-fold dynamic ambulance allocation problem, including forecasting the distribution of Emergency Medical Service (EMS) requesters and dynamically allocating ambulances according to the predicted distribution of requesters. In the definition of the coverage region, the Euclidean distance was used, which is not suitable for measuring the length of a route between two places. This study improved on the previous one by redefining the coverage region for practical application and providing a simulation model to verify the effectiveness of the proposed ambulance allocation method. The simulation results show the proposed allocation method providing higher demand coverage rates and shorter response distances than the official allocation.
在之前的研究中,我们解决了双重动态救护车分配问题,包括预测紧急医疗服务(EMS)请求者的分布,并根据预测的请求者分布动态分配救护车。在覆盖区域的定义中,使用欧几里得距离,不适合测量两地之间的路线长度。本研究在前人的基础上进行了改进,重新定义了实际应用的覆盖区域,并提供了仿真模型来验证所提出的救护车分配方法的有效性。仿真结果表明,该分配方法比官方分配方法具有更高的需求覆盖率和更短的响应距离。
{"title":"Empirical Spatial Density-Based Emergency Medical Service Demand Forecast for Ambulance Allocation","authors":"Y. Tsai, Hwei-Jen Lin, Pei Chi, Kelvin W. Lee","doi":"10.1142/S2424922X21500030","DOIUrl":"https://doi.org/10.1142/S2424922X21500030","url":null,"abstract":"In a previous study, we solved the two-fold dynamic ambulance allocation problem, including forecasting the distribution of Emergency Medical Service (EMS) requesters and dynamically allocating ambulances according to the predicted distribution of requesters. In the definition of the coverage region, the Euclidean distance was used, which is not suitable for measuring the length of a route between two places. This study improved on the previous one by redefining the coverage region for practical application and providing a simulation model to verify the effectiveness of the proposed ambulance allocation method. The simulation results show the proposed allocation method providing higher demand coverage rates and shorter response distances than the official allocation.","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"31 1","pages":"2150003:1-2150003:23"},"PeriodicalIF":0.6,"publicationDate":"2021-02-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80079864","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Editorial: Special Issue on New Frontiers in Data Sciences and Data Analytics Tools and Applications 社论:关于数据科学和数据分析工具与应用新领域的特刊
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2020-12-08 DOI: 10.1142/s2424922x20020015
Asadullah Shaikh, Khairan D. Rajab
{"title":"Editorial: Special Issue on New Frontiers in Data Sciences and Data Analytics Tools and Applications","authors":"Asadullah Shaikh, Khairan D. Rajab","doi":"10.1142/s2424922x20020015","DOIUrl":"https://doi.org/10.1142/s2424922x20020015","url":null,"abstract":"","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"133 1","pages":"2002001:1-2002001:3"},"PeriodicalIF":0.6,"publicationDate":"2020-12-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72825105","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Approximate Entropy and Empirical Mode Decomposition for Improved Speaker Recognition 改进说话人识别的近似熵和经验模态分解
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2020-12-02 DOI: 10.1142/s2424922x20500114
R. A. Metzger, J. Doherty, D. Jenkins, D. L. Hall
When processing real-world recordings of speech, it is highly probable noise will be present at some instance in the signal. Compounding this problem is the situation when the noise occurs in short, impulsive bursts at random intervals. Traditional signal processing methods used to detect speech rely on the spectral energy of the incoming signal to make a determination whether or not a segment of the signal contains speech. However when noise is present, this simple energy detection is prone to falsely flagging noise as speech. This paper will demonstrate an alternative way of processing a noisy speech signal utilizing a combination of information theoretic and signal processing principles to differentiate speech segments from noise. The utilization of this preprocessing technique will allow a speaker recognition system to train statistical speaker model using noise-corrupted speech files, and construct models statistically similar to those constructed from noise-free data. This preprocessing method will be shown to outperform traditional spectrum-based methods for both low-entropy and high-entropy noise in low signal-to-noise ratio environments, with a reduction in the feature space distortion when measured using the Cauchy–Schwarz (CS) distance metric.
在处理真实的语音记录时,很可能在信号的某些实例中存在噪声。使这个问题更加复杂的是,噪声以随机间隔的短脉冲爆发的情况。用于检测语音的传统信号处理方法依赖于输入信号的频谱能量来确定信号的一段是否包含语音。然而,当噪声存在时,这种简单的能量检测容易错误地将噪声标记为语音。本文将展示一种处理噪声语音信号的替代方法,利用信息论和信号处理原理的结合来区分语音片段和噪声。利用这种预处理技术,说话人识别系统可以使用被噪声破坏的语音文件来训练统计说话人模型,并构建与无噪声数据相似的统计模型。在低信噪比环境中,这种预处理方法将优于传统的基于频谱的低熵和高熵噪声方法,并且在使用Cauchy-Schwarz (CS)距离度量测量时减少了特征空间失真。
{"title":"Approximate Entropy and Empirical Mode Decomposition for Improved Speaker Recognition","authors":"R. A. Metzger, J. Doherty, D. Jenkins, D. L. Hall","doi":"10.1142/s2424922x20500114","DOIUrl":"https://doi.org/10.1142/s2424922x20500114","url":null,"abstract":"When processing real-world recordings of speech, it is highly probable noise will be present at some instance in the signal. Compounding this problem is the situation when the noise occurs in short, impulsive bursts at random intervals. Traditional signal processing methods used to detect speech rely on the spectral energy of the incoming signal to make a determination whether or not a segment of the signal contains speech. However when noise is present, this simple energy detection is prone to falsely flagging noise as speech. This paper will demonstrate an alternative way of processing a noisy speech signal utilizing a combination of information theoretic and signal processing principles to differentiate speech segments from noise. The utilization of this preprocessing technique will allow a speaker recognition system to train statistical speaker model using noise-corrupted speech files, and construct models statistically similar to those constructed from noise-free data. This preprocessing method will be shown to outperform traditional spectrum-based methods for both low-entropy and high-entropy noise in low signal-to-noise ratio environments, with a reduction in the feature space distortion when measured using the Cauchy–Schwarz (CS) distance metric.","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"20 1","pages":"2050011:1-2050011:24"},"PeriodicalIF":0.6,"publicationDate":"2020-12-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77969513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On Pricing American Put Option on a Fixed Term: A Monte Carlo Approach 固定期限美式看跌期权定价:蒙特卡洛方法
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2020-11-30 DOI: 10.1142/s2424922x20500102
Perpetual Andam Boiquaye
This paper focuses primarily on pricing an American put option with a fixed term where the price process is geometric mean-reverting. The change of measure is assumed to be incorporated. Monte Carlo simulation was used to calculate the price of the option and the results obtained were analyzed. The option price was found to be $94.42 and the optimal stopping time was approximately one year after the option was sold which means that exercising early is the best for an American put option on a fixed term. Also, the seller of the put option should have sold $0.01 assets and bought $95.51 bonds to get the same payoff as the buyer at the end of one year for it to be a zero-sum game. In the simulation study, the parameters were varied to see the influence it had on the option price and the stopping time and it showed that it either increases or decreases the value of the option price and the optimal stopping time or it remained unchanged.
本文主要研究具有固定期限的美式看跌期权的定价过程,其价格过程是几何均值回归的。度量的变化被认为是合并的。采用蒙特卡罗模拟方法对期权价格进行了计算,并对计算结果进行了分析。期权价格为94.42美元,最优止损时间约为期权卖出后一年,这意味着对于固定期限美式看跌期权,提前行使是最好的。此外,看跌期权的卖方应该卖出0.01美元的资产,买入95.51美元的债券,以便在一年后获得与买方相同的收益,因为这是一个零和游戏。在模拟研究中,通过改变参数来观察其对期权价格和最优停止时间的影响,结果表明,该参数或增加或减少期权价格和最优停止时间的值,或保持不变。
{"title":"On Pricing American Put Option on a Fixed Term: A Monte Carlo Approach","authors":"Perpetual Andam Boiquaye","doi":"10.1142/s2424922x20500102","DOIUrl":"https://doi.org/10.1142/s2424922x20500102","url":null,"abstract":"This paper focuses primarily on pricing an American put option with a fixed term where the price process is geometric mean-reverting. The change of measure is assumed to be incorporated. Monte Carlo simulation was used to calculate the price of the option and the results obtained were analyzed. The option price was found to be $94.42 and the optimal stopping time was approximately one year after the option was sold which means that exercising early is the best for an American put option on a fixed term. Also, the seller of the put option should have sold $0.01 assets and bought $95.51 bonds to get the same payoff as the buyer at the end of one year for it to be a zero-sum game. In the simulation study, the parameters were varied to see the influence it had on the option price and the stopping time and it showed that it either increases or decreases the value of the option price and the optimal stopping time or it remained unchanged.","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"54 1","pages":"2050010:1-2050010:11"},"PeriodicalIF":0.6,"publicationDate":"2020-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77146997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Features Reweighting and Selection in ligand-based Virtual Screening for Molecular Similarity Searching Based on Deep Belief Networks 基于深度信念网络的分子相似性搜索配体虚拟筛选中的特征重加权与选择
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2020-11-30 DOI: 10.1142/s2424922x20500096
Maged Nasser, N. Salim, Hamza Hentabli, Faisal Saeed, I. Rabiu
Virtual screening (VS) is defined as the use of a compilation of computational procedures to grade, score and/or sort several chemical formations. The purpose of VS is to identify the molecules holding the greatest prior probabilities of activity. Many of the conventional similarity methods assume that molecular features that do not relate to the biological activity carry the same weight as the important ones. For this reason, the researchers on this paper investigated that some features are being more important than others through the chemist structure diagrams and the weight for each fragment should be taken into consideration by giving more weight to those fragments that are more important. In this paper, a deep learning method specifically known as Deep Belief Networks (DBN) has been used to reweight the molecule features and based on this new weigh, the reconstruction feature error has been calculated for all the features. Based on the reconstruction feature error values, Principal Component Analysis (PCA) has been used for the dimension’s reduction and only few hundreds of features have been selected based on the less error rate. The main aim of this research is to show an improvement of the similarity searching performance based on the selected features those have less error rate. The results derived through the DBN were compared with those derived through other similarity methods, such as the Tanimoto coefficient and the quantum-based methods. This comparison revealed the performance of the DBN with the structurally heterogeneous data sets (DS1 and DS3) to be superior to the performances of all the other techniques.
虚拟筛选(VS)被定义为使用一系列计算程序对几种化学地层进行分级、评分和/或分类。VS的目的是识别具有最大活动先验概率的分子。许多传统的相似性方法假设与生物活性无关的分子特征与重要的分子特征具有相同的权重。因此,本文研究人员通过化学结构图研究了某些特征比其他特征更重要,并且应该考虑每个片段的权重,给予更重要的片段更多的权重。本文采用深度学习方法深度信念网络(deep Belief Networks, DBN)对分子特征进行重加权,并在此基础上计算所有特征的重构特征误差。基于重构特征的误差值,采用主成分分析(PCA)进行降维,在错误率较小的基础上只选择了几百个特征。本研究的主要目的是通过选择错误率较小的特征来提高相似度搜索的性能。通过DBN得到的结果与其他相似方法得到的结果进行了比较,如谷本系数和基于量子的方法。这一比较揭示了DBN与结构异构数据集(DS1和DS3)的性能优于所有其他技术的性能。
{"title":"Features Reweighting and Selection in ligand-based Virtual Screening for Molecular Similarity Searching Based on Deep Belief Networks","authors":"Maged Nasser, N. Salim, Hamza Hentabli, Faisal Saeed, I. Rabiu","doi":"10.1142/s2424922x20500096","DOIUrl":"https://doi.org/10.1142/s2424922x20500096","url":null,"abstract":"Virtual screening (VS) is defined as the use of a compilation of computational procedures to grade, score and/or sort several chemical formations. The purpose of VS is to identify the molecules holding the greatest prior probabilities of activity. Many of the conventional similarity methods assume that molecular features that do not relate to the biological activity carry the same weight as the important ones. For this reason, the researchers on this paper investigated that some features are being more important than others through the chemist structure diagrams and the weight for each fragment should be taken into consideration by giving more weight to those fragments that are more important. In this paper, a deep learning method specifically known as Deep Belief Networks (DBN) has been used to reweight the molecule features and based on this new weigh, the reconstruction feature error has been calculated for all the features. Based on the reconstruction feature error values, Principal Component Analysis (PCA) has been used for the dimension’s reduction and only few hundreds of features have been selected based on the less error rate. The main aim of this research is to show an improvement of the similarity searching performance based on the selected features those have less error rate. The results derived through the DBN were compared with those derived through other similarity methods, such as the Tanimoto coefficient and the quantum-based methods. This comparison revealed the performance of the DBN with the structurally heterogeneous data sets (DS1 and DS3) to be superior to the performances of all the other techniques.","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"26 1","pages":"2050009:1-2050009:28"},"PeriodicalIF":0.6,"publicationDate":"2020-11-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78130310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Towards a Knowledge-Driven Framework of Cyber-Physical Social System for Multidimensional Urban Mobility 面向多维城市交通的信息物理社会系统知识驱动框架
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2020-11-26 DOI: 10.1142/s2424922x20410053
F. Abbasi, Mohamed-Hedi Karray, R. Houé, M. Memon, B. Archimède
This paper focuses on the proposed generic framework of Cyber-Physical Social Systems (CPSS) for Multidimensional Urban Mobility (MUM). The introduction of air mobility will increase the complexity of urban mobility management due to the increase of data (real-time information exchange), services, and infrastructure. This paper first summarizes the shift of the smart mobility and associated concerns and emphasizes the smart technologies to be introduced for efficient deployment of the MUM. Afterwards, the state-of-the-art on CPSS present the details of the research work in the context of CPSS data management process and challenges related with data. In conclusion, a framework to address the various challenges associated with the data paradigm is proposed.
本文重点研究了面向多维城市交通(MUM)的信息物理社会系统(CPSS)的通用框架。由于数据(实时信息交换)、服务和基础设施的增加,空中交通的引入将增加城市交通管理的复杂性。本文首先总结了智能交通的转变及其相关问题,并强调了为有效部署智能交通需要引入的智能技术。随后,介绍了CPSS数据管理过程中研究工作的细节以及与数据相关的挑战。最后,提出了一个框架来解决与数据范式相关的各种挑战。
{"title":"Towards a Knowledge-Driven Framework of Cyber-Physical Social System for Multidimensional Urban Mobility","authors":"F. Abbasi, Mohamed-Hedi Karray, R. Houé, M. Memon, B. Archimède","doi":"10.1142/s2424922x20410053","DOIUrl":"https://doi.org/10.1142/s2424922x20410053","url":null,"abstract":"This paper focuses on the proposed generic framework of Cyber-Physical Social Systems (CPSS) for Multidimensional Urban Mobility (MUM). The introduction of air mobility will increase the complexity of urban mobility management due to the increase of data (real-time information exchange), services, and infrastructure. This paper first summarizes the shift of the smart mobility and associated concerns and emphasizes the smart technologies to be introduced for efficient deployment of the MUM. Afterwards, the state-of-the-art on CPSS present the details of the research work in the context of CPSS data management process and challenges related with data. In conclusion, a framework to address the various challenges associated with the data paradigm is proposed.","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"63 1","pages":"2041005:1-2041005:11"},"PeriodicalIF":0.6,"publicationDate":"2020-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80179912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Smartphone-Based Pavement Roughness Estimation Using Deep Learning with Entity Embedding 基于实体嵌入深度学习的智能手机路面粗糙度估计
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2020-11-26 DOI: 10.1142/s2424922x20500072
Armstrong Aboah, Y. Adu-Gyamfi
The commonly used index for measuring pavement roughness is the International Roughness index (IRI). Traditional method for collecting road surface information is expensive and as such researchers over the years have resorted to other cheaper ways of collecting data. This study focuses on developing a deep learning model to quickly and accurately determine the IRI values of road sections at a cheaper cost. The study proposed a model that uses accelerometer data and previous year’s IRI values to predict current year IRI values. The study concludes that addition of accelerometer readings to previous year’s IRIs increased the accuracy of prediction.
衡量路面粗糙度的常用指标是国际粗糙度指数(IRI)。收集路面信息的传统方法是昂贵的,因此研究人员多年来一直采用其他更便宜的收集数据的方法。本研究的重点是开发一种深度学习模型,以更低的成本快速准确地确定路段的IRI值。该研究提出了一个模型,使用加速度计数据和前一年的IRI值来预测当前年份的IRI值。该研究的结论是,在前一年的IRIs中增加加速度计的读数提高了预测的准确性。
{"title":"Smartphone-Based Pavement Roughness Estimation Using Deep Learning with Entity Embedding","authors":"Armstrong Aboah, Y. Adu-Gyamfi","doi":"10.1142/s2424922x20500072","DOIUrl":"https://doi.org/10.1142/s2424922x20500072","url":null,"abstract":"The commonly used index for measuring pavement roughness is the International Roughness index (IRI). Traditional method for collecting road surface information is expensive and as such researchers over the years have resorted to other cheaper ways of collecting data. This study focuses on developing a deep learning model to quickly and accurately determine the IRI values of road sections at a cheaper cost. The study proposed a model that uses accelerometer data and previous year’s IRI values to predict current year IRI values. The study concludes that addition of accelerometer readings to previous year’s IRIs increased the accuracy of prediction.","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"37 1","pages":"2050007:1-2050007:20"},"PeriodicalIF":0.6,"publicationDate":"2020-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90218408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Roman Urdu Headline News Text Classification Using RNN, LSTM and CNN 罗马乌尔都语标题新闻文本分类使用RNN, LSTM和CNN
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2020-11-18 DOI: 10.1142/s2424922x20500084
Irfan Ali Kandhro, Sahar Zafar Jumani, K. Kumar, Abdul Hafeez, F. Ali
This paper presents the automated tool for the classification of text with respect to predefined categories. It has always been considered as a vital method to manage and process a huge number of documents in digital forms which are widespread and continuously increasing. Most of the research work in text classification has been done in Urdu, English and other languages. But limited research work has been carried out on roman data. Technically, the process of the text classification follows two steps: the first step consists of choosing the main features from all the available features of the text documents with the usage of feature extraction techniques. The second step applies classification algorithms on those chosen features. The data set is collected through scraping tools from the most popular news websites Awaji Awaze and Daily Jhoongar. Furthermore, the data set splits in training and testing 70%, 30%, respectively. In this paper, the deep learning models, such as RNN, LSTM, and CNN, are used for classification of roman Urdu headline news. The testing accuracy of RNN (81%), LSTM (82%), and CNN (79%), and the experimental results demonstrate that the performance of the LSTM method is state-of-art method compared to CNN and RNN.
本文提出了一种基于预定义分类的文本自动分类工具。它一直被认为是管理和处理数字形式的大量文件的重要方法,这些文件广泛存在并不断增加。大多数文本分类的研究工作都是在乌尔都语、英语和其他语言中完成的。但是,对罗马资料的研究工作有限。从技术上讲,文本分类的过程分为两个步骤:第一步是使用特征提取技术从文本文档的所有可用特征中选择主要特征。第二步对这些选择的特征应用分类算法。数据集是通过最受欢迎的新闻网站Awaji Awaze和Daily Jhoongar的抓取工具收集的。此外,数据集在训练和测试中分别分裂70%和30%。本文采用RNN、LSTM、CNN等深度学习模型对罗马乌尔都语标题新闻进行分类。RNN(81%)、LSTM(82%)和CNN(79%)的测试准确率,实验结果表明LSTM方法的性能与CNN和RNN相比是最先进的方法。
{"title":"Roman Urdu Headline News Text Classification Using RNN, LSTM and CNN","authors":"Irfan Ali Kandhro, Sahar Zafar Jumani, K. Kumar, Abdul Hafeez, F. Ali","doi":"10.1142/s2424922x20500084","DOIUrl":"https://doi.org/10.1142/s2424922x20500084","url":null,"abstract":"This paper presents the automated tool for the classification of text with respect to predefined categories. It has always been considered as a vital method to manage and process a huge number of documents in digital forms which are widespread and continuously increasing. Most of the research work in text classification has been done in Urdu, English and other languages. But limited research work has been carried out on roman data. Technically, the process of the text classification follows two steps: the first step consists of choosing the main features from all the available features of the text documents with the usage of feature extraction techniques. The second step applies classification algorithms on those chosen features. The data set is collected through scraping tools from the most popular news websites Awaji Awaze and Daily Jhoongar. Furthermore, the data set splits in training and testing 70%, 30%, respectively. In this paper, the deep learning models, such as RNN, LSTM, and CNN, are used for classification of roman Urdu headline news. The testing accuracy of RNN (81%), LSTM (82%), and CNN (79%), and the experimental results demonstrate that the performance of the LSTM method is state-of-art method compared to CNN and RNN.","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"87 1","pages":"2050008:1-2050008:13"},"PeriodicalIF":0.6,"publicationDate":"2020-11-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80844137","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Continuous Empirical Wavelets Systems 连续经验小波系统
IF 0.6 Q4 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Pub Date : 2020-11-17 DOI: 10.1142/S2424922X20500060
J. Gilles
The recently proposed empirical wavelet transform was based on a particular type of filter. In this paper, we aim to propose a general framework for the construction of empirical wavelet systems in the continuous case. We define a well-suited formalism and then investigate some general properties of empirical wavelet systems. In particular, we provide some sufficient conditions to the existence of a reconstruction formula. In the second part of the paper, we propose the construction of empirical wavelet systems based on some classic mother wavelets.
最近提出的经验小波变换是基于一种特殊类型的滤波器。在本文中,我们的目的是提出一个构造连续情况下经验小波系统的一般框架。我们定义了一个合适的形式,然后研究了经验小波系统的一些一般性质。特别地,我们给出了重构公式存在的一些充分条件。在论文的第二部分,我们在一些经典母小波的基础上提出了经验小波系统的构造。
{"title":"Continuous Empirical Wavelets Systems","authors":"J. Gilles","doi":"10.1142/S2424922X20500060","DOIUrl":"https://doi.org/10.1142/S2424922X20500060","url":null,"abstract":"The recently proposed empirical wavelet transform was based on a particular type of filter. In this paper, we aim to propose a general framework for the construction of empirical wavelet systems in the continuous case. We define a well-suited formalism and then investigate some general properties of empirical wavelet systems. In particular, we provide some sufficient conditions to the existence of a reconstruction formula. In the second part of the paper, we propose the construction of empirical wavelet systems based on some classic mother wavelets.","PeriodicalId":47145,"journal":{"name":"Advances in Data Science and Adaptive Analysis","volume":"48 1","pages":"2050006:1-2050006:21"},"PeriodicalIF":0.6,"publicationDate":"2020-11-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87936350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Advances in Data Science and Adaptive Analysis
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1