首页 > 最新文献

Annals of Data Science最新文献

英文 中文
Bayesian Analysis of Generalized Inverted Exponential Distribution Based on Generalized Progressive Hybrid Censoring Competing Risks Data 基于竞争风险数据的广义逆指数分布的贝叶斯分析
Q1 Decision Sciences Pub Date : 2023-08-21 DOI: 10.1007/s40745-023-00488-y
Amal S. Hassan, Rana M. Mousa, Mahmoud H. Abu-Moussa

In this study, a competing risk model was developed under a generalized progressive hybrid censoring scheme using a generalized inverted exponential distribution. The latent causes of failure were presumed to be independent. Estimating the unknown parameters is performed using maximum likelihood (ML) and Bayesian methods. Using the Markov chain Monte Carlo technique, Bayesian estimators were obtained under gamma priors with various loss functions. ML estimate was used to create confidence intervals (CIs). In addition, we present two bootstrap CIs for the unknown parameters. Further, credible CIs and the highest posterior density intervals were constructed based on the conditional posterior distribution. Monte Carlo simulation is used to examine the performance of different estimates. Applications to real data were used to check the estimates and compare the proposed model with alternative distributions.

本研究采用广义倒指数分布,在广义渐进混合删减方案下建立了竞争风险模型。假定故障的潜在原因是独立的。使用最大似然法(ML)和贝叶斯法估计未知参数。使用马尔科夫链蒙特卡洛技术,在伽马先验条件下通过各种损失函数获得贝叶斯估计值。ML估计值用于创建置信区间(CI)。此外,我们还提出了两个未知参数的引导置信区间。此外,我们还根据条件后验分布构建了可信 CI 和最高后验密度区间。蒙特卡罗模拟用于检验不同估计值的性能。对真实数据的应用被用来检查估计值,并将提出的模型与其他分布进行比较。
{"title":"Bayesian Analysis of Generalized Inverted Exponential Distribution Based on Generalized Progressive Hybrid Censoring Competing Risks Data","authors":"Amal S. Hassan,&nbsp;Rana M. Mousa,&nbsp;Mahmoud H. Abu-Moussa","doi":"10.1007/s40745-023-00488-y","DOIUrl":"10.1007/s40745-023-00488-y","url":null,"abstract":"<div><p>In this study, a competing risk model was developed under a generalized progressive hybrid censoring scheme using a generalized inverted exponential distribution. The latent causes of failure were presumed to be independent. Estimating the unknown parameters is performed using maximum likelihood (ML) and Bayesian methods. Using the Markov chain Monte Carlo technique, Bayesian estimators were obtained under gamma priors with various loss functions. ML estimate was used to create confidence intervals (CIs). In addition, we present two bootstrap CIs for the unknown parameters. Further, credible CIs and the highest posterior density intervals were constructed based on the conditional posterior distribution. Monte Carlo simulation is used to examine the performance of different estimates. Applications to real data were used to check the estimates and compare the proposed model with alternative distributions.\u0000</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 4","pages":"1225 - 1264"},"PeriodicalIF":0.0,"publicationDate":"2023-08-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48748310","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Quantitative Analysis of Group for Epidemiology Architectural Approach 流行病学体系结构方法的群体定量分析
Q1 Decision Sciences Pub Date : 2023-08-18 DOI: 10.1007/s40745-023-00493-1
Dephney Mathebula

Epidemiology, the aspect of research focusing on disease modelling is date intensive. Research epidemiologists in different research groups played a key role in developing different data driven model for COVID-19 and monkeypox. The requirement of accessing highly accurate data useful for disease modelling is beneficial but not without having challenges. Currently, the task of data acquisition is executed by select individuals in different research groups. This approach experiences the drawbacks associated with getting permission to access the desired data and inflexibility to change data acquisition goals due to dynamic epidemiological research objectives. The presented research addresses these challenges and proposes the design and use of dynamic intelligent crawlers for acquiring epidemiological data related to a given goal. In addition, the research aims to quantify how the use of computing entities enhances the process of data acquisition in epidemiological related studies. This is done by formulating and investigating the metrics of the data acquisition efficiency and the data analytics efficiency. The use of human assisted crawlers in the global information networks is found to enhance data acquisition efficiency (DAqE) and data analytics efficiency (DAnE). The use of human assisted crawlers in a hybrid configuration outperforms the case where manual research group member efforts are expended enhancing the DAqE and DAnE by up to 35% and 99% on average, respectively.

流行病学是以疾病建模为重点的研究领域,需要大量数据。不同研究小组的流行病学研究人员在为 COVID-19 和猴痘开发不同的数据驱动模型方面发挥了关键作用。获取对疾病建模有用的高精度数据的要求是有益的,但也并非没有挑战。目前,获取数据的任务由不同研究小组的选定人员执行。这种方法的缺点是需要获得访问所需数据的许可,而且由于流行病学研究目标的不断变化,数据采集目标的改变也不灵活。本研究针对这些挑战,提出了设计和使用动态智能爬虫来获取与给定目标相关的流行病学数据的建议。此外,研究还旨在量化计算实体的使用如何增强流行病学相关研究的数据采集过程。具体做法是制定和研究数据获取效率和数据分析效率的衡量标准。研究发现,在全球信息网络中使用人工辅助爬虫可提高数据获取效率(DAqE)和数据分析效率(DAnE)。在混合配置中使用人工辅助爬虫的效果优于人工研究小组成员工作的情况,平均可分别提高 35% 和 99% 的 DAqE 和 DAnE。
{"title":"Quantitative Analysis of Group for Epidemiology Architectural Approach","authors":"Dephney Mathebula","doi":"10.1007/s40745-023-00493-1","DOIUrl":"10.1007/s40745-023-00493-1","url":null,"abstract":"<div><p>Epidemiology, the aspect of research focusing on disease modelling is date intensive. Research epidemiologists in different research groups played a key role in developing different data driven model for COVID-19 and monkeypox. The requirement of accessing highly accurate data useful for disease modelling is beneficial but not without having challenges. Currently, the task of data acquisition is executed by select individuals in different research groups. This approach experiences the drawbacks associated with getting permission to access the desired data and inflexibility to change data acquisition goals due to dynamic epidemiological research objectives. The presented research addresses these challenges and proposes the design and use of dynamic intelligent crawlers for acquiring epidemiological data related to a given goal. In addition, the research aims to quantify how the use of computing entities enhances the process of data acquisition in epidemiological related studies. This is done by formulating and investigating the metrics of the data acquisition efficiency and the data analytics efficiency. The use of human assisted crawlers in the global information networks is found to enhance data acquisition efficiency (DAqE) and data analytics efficiency (DAnE). The use of human assisted crawlers in a hybrid configuration outperforms the case where manual research group member efforts are expended enhancing the DAqE and DAnE by up to 35% and 99% on average, respectively.</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 3","pages":"979 - 1001"},"PeriodicalIF":0.0,"publicationDate":"2023-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s40745-023-00493-1.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"44922195","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Estimation of ( P[Y 具有威布尔边际的应力-强度模型依赖性的$$P[Y<X]$$估计
Q1 Decision Sciences Pub Date : 2023-08-17 DOI: 10.1007/s40745-023-00487-z
Dipak D. Patil, U. V. Naik-Nimbalkar, M. M. Kale

The stress–strength model is a basic tool used in evaluating the reliability ( R = P(Y < X)). We consider an expression for R where the random variables X and Y denote strength and stress, respectively. The system fails only if the stress exceeds the strength. We aim to study the effect of the dependency between X and Y on R. We assume that X and Y follow Weibull distributions and their dependency is modeled by a copula with the dependency parameter ( theta ). We compute R for Farlie–Gumbel–Morgenstern (FGM), Ali–Mikhail–Haq (AMH), Gumbel’s bivariate exponential copulas, and for Gumbel–Hougaard (GH) copula using a Monte-Carlo integration technique. We plot the graph of R versus (theta ) to study the effect of dependency on R. We estimate R by plugging in the estimates of the marginal parameters and of ( theta ) in its expression. The estimates of the marginal parameters are based on the marginal likelihood. The estimates of (theta ) are obtained from two different methods; one is based on the conditional likelihood and the other is based on the method of moments using Blomqvist’s beta. Asymptotic distribution of both the estimators of R is obtained. Finally, analysis of real data set is also performed for illustrative purposes.

应力-强度模型是用于评估可靠性的基本工具(R = P(Y < X))。我们考虑 R 的表达式,其中随机变量 X 和 Y 分别表示强度和应力。只有当应力超过强度时,系统才会失效。我们假定 X 和 Y 遵循 Weibull 分布,它们之间的依赖关系由具有依赖参数 ( theta )的 copula 来建模。我们使用蒙特卡洛积分技术计算 Farlie-Gumbel-Morgenstern (FGM)、Ali-Mikhail-Haq (AMH)、Gumbel 的双变量指数协程以及 Gumbel-Hougaard (GH) 协程的 R。我们绘制了 R 与 (theta )的关系图,以研究依赖性对 R 的影响。我们在 R 的表达式中插入边际参数和 (theta )的估计值来估计 R。边际参数的估计基于边际似然法。(theta )的估计值由两种不同的方法获得;一种是基于条件似然法,另一种是基于使用布隆奎斯特贝塔的矩方法。两种 R 估计数的渐近分布都已得到。最后,还对真实数据集进行了分析,以说明问题。
{"title":"Estimation of ( P[Y<X] ) for Dependence of Stress–Strength Models with Weibull Marginals","authors":"Dipak D. Patil,&nbsp;U. V. Naik-Nimbalkar,&nbsp;M. M. Kale","doi":"10.1007/s40745-023-00487-z","DOIUrl":"10.1007/s40745-023-00487-z","url":null,"abstract":"<div><p>The stress–strength model is a basic tool used in evaluating the reliability <span>( R = P(Y &lt; X))</span>. We consider an expression for <i>R</i> where the random variables X and Y denote strength and stress, respectively. The system fails only if the stress exceeds the strength. We aim to study the effect of the dependency between X and Y on <i>R</i>. We assume that X and Y follow Weibull distributions and their dependency is modeled by a copula with the dependency parameter <span>( theta )</span>. We compute <i>R</i> for Farlie–Gumbel–Morgenstern (FGM), Ali–Mikhail–Haq (AMH), Gumbel’s bivariate exponential copulas, and for Gumbel–Hougaard (GH) copula using a Monte-Carlo integration technique. We plot the graph of <i>R</i> versus <span>(theta )</span> to study the effect of dependency on <i>R</i>. We estimate <i>R</i> by plugging in the estimates of the marginal parameters and of <span>( theta )</span> in its expression. The estimates of the marginal parameters are based on the marginal likelihood. The estimates of <span>(theta )</span> are obtained from two different methods; one is based on the conditional likelihood and the other is based on the method of moments using Blomqvist’s beta. Asymptotic distribution of both the estimators of <i>R</i> is obtained. Finally, analysis of real data set is also performed for illustrative purposes.</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 4","pages":"1303 - 1340"},"PeriodicalIF":0.0,"publicationDate":"2023-08-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48578734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Modi-Weibull Distribution: Inferential and Simulation Study 莫迪-威布尔分布:推理与仿真研究
Q1 Decision Sciences Pub Date : 2023-08-15 DOI: 10.1007/s40745-023-00491-3
Harshita Kumawat, Kanak Modi, Pankaj Nagar

This paper presents a study on a new family of distributions using the Weibull distribution and termed as Modi-Weibull distribution. This Modi-Weibull distribution is based on four parameters. To understand the behaviour of the distribution, some statistical characteristics have been derived, such as shapes of density and distribution function, hazard function, survival function, median, moments, order statistics etc. These parameters are estimated using classical maximum likelihood estimation method. Asymptotic confidence intervals for parameters of Modi-Weibull distribution are also obtained. A simulation study is carried out to investigate the bias, MSE of proposed maximum likelihood estimators along with coverage probability and average width of confidence intervals of parameters. Two applications to real data sets are discussed to illustrate the fitting of the proposed distribution and compared with some well-known distributions.

本文研究了一种使用威布尔分布的新分布族,称为莫迪-威布尔分布。莫迪-韦布尔分布基于四个参数。为了解该分布的行为,我们推导出了一些统计特征,如密度和分布函数的形状、危险函数、生存函数、中位数、矩、阶次统计等。这些参数采用经典的最大似然估计法进行估计。此外,还获得了莫迪-韦布尔分布参数的渐近置信区间。我们还进行了模拟研究,以调查所提出的最大似然估计方法的偏差、MSE 以及参数置信区间的覆盖概率和平均宽度。讨论了两个真实数据集的应用,以说明拟议分布的拟合情况,并与一些著名的分布进行了比较。
{"title":"Modi-Weibull Distribution: Inferential and Simulation Study","authors":"Harshita Kumawat,&nbsp;Kanak Modi,&nbsp;Pankaj Nagar","doi":"10.1007/s40745-023-00491-3","DOIUrl":"10.1007/s40745-023-00491-3","url":null,"abstract":"<div><p>This paper presents a study on a new family of distributions using the Weibull distribution and termed as Modi-Weibull distribution. This Modi-Weibull distribution is based on four parameters. To understand the behaviour of the distribution, some statistical characteristics have been derived, such as shapes of density and distribution function, hazard function, survival function, median, moments, order statistics etc. These parameters are estimated using classical maximum likelihood estimation method. Asymptotic confidence intervals for parameters of Modi-Weibull distribution are also obtained. A simulation study is carried out to investigate the bias, MSE of proposed maximum likelihood estimators along with coverage probability and average width of confidence intervals of parameters. Two applications to real data sets are discussed to illustrate the fitting of the proposed distribution and compared with some well-known distributions.</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 6","pages":"1975 - 1999"},"PeriodicalIF":0.0,"publicationDate":"2023-08-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s40745-023-00491-3.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"48772839","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Shrinkage Estimation for Location and Scale Parameters of Logistic Distribution Under Record Values 记录值下物流配送位置和规模参数的收缩估计
Q1 Decision Sciences Pub Date : 2023-08-14 DOI: 10.1007/s40745-023-00492-2
Shubham Gupta, Gajendra K. Vishwakarma, A. M. Elsawah

Logistic distribution (LogDis) is frequently used in many different applications, such as logistic regression, logit models, classification, neural networks, physical sciences, sports modeling, finance and health and disease studies. For instance, the distribution function of the LogDis has the same functional form as the derivative of the Fermi function that can be used to set the relative weight of various electron energies in their contributions to electron transport. The LogDis has wider tails than a normal distribution (NorDis), so it is more consistent with the underlying data and provides better insight into the likelihood of extreme events. For this reason the United States Chess Federation has switched its formula for calculating chess ratings from the NorDis to the LogDis. The outcomes of many real-life experiments are sequences of record-breaking data sets, where only observations that exceed (or only those that fall below) the current extreme value are recorded. The practice demonstrated that the widely used estimators of the scale and location parameters of logistic record values, such as the best linear unbiased estimators (BLUEs), have some defects. This paper investigates the shrinkage estimators of the location and scale parameters for logistic record values using prior information about their BLUEs. Theoretical and computational justifications for the accuracy and precision of the proposed shrinkage estimators are investigated via their bias and mean square error (MSE), which provide sufficient conditions for improving the proposed shrinkage estimators to get unbiased estimators with minimum MSE. The performance of the proposed shrinkage estimators is compared with the performances of the BLUEs. The results demonstrate that the resulting shrinkage estimators are shown to be remarkably efficient.

逻辑分布(LogDis)经常被用于许多不同的应用中,如逻辑回归、Logit 模型、分类、神经网络、物理科学、体育建模、金融以及健康和疾病研究。例如,LogDis 的分布函数与费米函数的导数具有相同的函数形式,可用于设定各种电子能量对电子传输贡献的相对权重。与正态分布(NorDis)相比,LogDis 的尾部更宽,因此与基础数据更一致,能更好地洞察极端事件发生的可能性。因此,美国国际象棋联合会已将国际象棋等级分的计算公式从 NorDis 改为 LogDis。许多现实生活中的实验结果都是破纪录的数据集序列,其中只有超过(或只有低于)当前极值的观测数据才会被记录下来。实践证明,广泛使用的对数记录值的尺度和位置参数估计器,如最佳线性无偏估计器(BLUEs),存在一些缺陷。本文利用逻辑记录值的最佳线性无偏估计值的先验信息,研究了逻辑记录值的位置和尺度参数的收缩估计值。通过偏差和均方误差(MSE)研究了所提出的收缩估计器的准确性和精确性的理论和计算理由,为改进所提出的收缩估计器以获得最小 MSE 的无偏估计器提供了充分条件。建议的收缩估计器的性能与 BLUE 的性能进行了比较。结果表明,所得到的收缩估计器非常高效。
{"title":"Shrinkage Estimation for Location and Scale Parameters of Logistic Distribution Under Record Values","authors":"Shubham Gupta,&nbsp;Gajendra K. Vishwakarma,&nbsp;A. M. Elsawah","doi":"10.1007/s40745-023-00492-2","DOIUrl":"10.1007/s40745-023-00492-2","url":null,"abstract":"<div><p>Logistic distribution (LogDis) is frequently used in many different applications, such as logistic regression, logit models, classification, neural networks, physical sciences, sports modeling, finance and health and disease studies. For instance, the distribution function of the LogDis has the same functional form as the derivative of the Fermi function that can be used to set the relative weight of various electron energies in their contributions to electron transport. The LogDis has wider tails than a normal distribution (NorDis), so it is more consistent with the underlying data and provides better insight into the likelihood of extreme events. For this reason the United States Chess Federation has switched its formula for calculating chess ratings from the NorDis to the LogDis. The outcomes of many real-life experiments are sequences of record-breaking data sets, where only observations that exceed (or only those that fall below) the current extreme value are recorded. The practice demonstrated that the widely used estimators of the scale and location parameters of logistic record values, such as the best linear unbiased estimators (BLUEs), have some defects. This paper investigates the shrinkage estimators of the location and scale parameters for logistic record values using prior information about their BLUEs. Theoretical and computational justifications for the accuracy and precision of the proposed shrinkage estimators are investigated via their bias and mean square error (MSE), which provide sufficient conditions for improving the proposed shrinkage estimators to get unbiased estimators with minimum MSE. The performance of the proposed shrinkage estimators is compared with the performances of the BLUEs. The results demonstrate that the resulting shrinkage estimators are shown to be remarkably efficient.</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 4","pages":"1209 - 1224"},"PeriodicalIF":0.0,"publicationDate":"2023-08-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46550442","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Bayesian Classifier Using Vine Copula and Fuzzy Clustering Technique 利用Vine Copula和模糊聚类技术改进贝叶斯分类器
Q1 Decision Sciences Pub Date : 2023-08-10 DOI: 10.1007/s40745-023-00490-4
Ha Che-Ngoc, Thao Nguyen-Trang, Hieu Huynh-Van, Tai Vo-Van

Classification is a fundamental problem in statistics and data science, and it has garnered significant interest from researchers. This research proposes a new classification algorithm that builds upon two key improvements of the Bayesian method. First, we introduce a method to determine the prior probabilities using fuzzy clustering techniques. The prior probability is determined based on the fuzzy level of the classified element within the groups. Second, we develop the probability density function using Vine Copula. By combining these improvements, we obtain an automatic classification algorithm with several advantages. The proposed algorithm is presented with specific steps and illustrated using numerical examples. Furthermore, it is applied to classify image data, demonstrating its significant potential in various real-world applications. The numerical examples and applications highlight that the proposed algorithm outperforms existing methods, including traditional statistics and machine learning approaches.

分类是统计学和数据科学中的一个基本问题,已引起研究人员的极大兴趣。本研究在贝叶斯方法的两个关键改进基础上提出了一种新的分类算法。首先,我们引入了一种利用模糊聚类技术确定先验概率的方法。先验概率是根据组内分类元素的模糊级别确定的。其次,我们使用 Vine Copula 开发了概率密度函数。结合这些改进,我们获得了一种具有多种优势的自动分类算法。我们通过具体步骤介绍了所提出的算法,并使用数字示例进行了说明。此外,该算法还被应用于图像数据分类,展示了其在各种实际应用中的巨大潜力。数字示例和应用突出表明,所提出的算法优于现有的方法,包括传统的统计和机器学习方法。
{"title":"Improving Bayesian Classifier Using Vine Copula and Fuzzy Clustering Technique","authors":"Ha Che-Ngoc,&nbsp;Thao Nguyen-Trang,&nbsp;Hieu Huynh-Van,&nbsp;Tai Vo-Van","doi":"10.1007/s40745-023-00490-4","DOIUrl":"10.1007/s40745-023-00490-4","url":null,"abstract":"<div><p>Classification is a fundamental problem in statistics and data science, and it has garnered significant interest from researchers. This research proposes a new classification algorithm that builds upon two key improvements of the Bayesian method. First, we introduce a method to determine the prior probabilities using fuzzy clustering techniques. The prior probability is determined based on the fuzzy level of the classified element within the groups. Second, we develop the probability density function using Vine Copula. By combining these improvements, we obtain an automatic classification algorithm with several advantages. The proposed algorithm is presented with specific steps and illustrated using numerical examples. Furthermore, it is applied to classify image data, demonstrating its significant potential in various real-world applications. The numerical examples and applications highlight that the proposed algorithm outperforms existing methods, including traditional statistics and machine learning approaches.</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 2","pages":"709 - 732"},"PeriodicalIF":0.0,"publicationDate":"2023-08-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"49453964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A New Lindley Extension: Estimation, Risk Assessment and Analysis Under Bimodal Right Skewed Precipitation Data 一个新的Lindley推广:双峰右偏降水数据下的估算、风险评估与分析
Q1 Decision Sciences Pub Date : 2023-08-08 DOI: 10.1007/s40745-023-00485-1
Majid Hashempour, Morad Alizadeh, Haitham M. Yousof

The objectives of this study are to propose a new two-parameter lifespan distribution and explain some of the most essential properties of that distribution. Through the course of this investigation, we will be able to achieve both of these objectives. For the aim of assessment, research is carried out that makes use of simulation, and for the same reason, a variety of various approaches are studied and taken into account for the purpose of evaluation. Making use of two separate data collections enables an analysis of the adaptability of the suggested distribution to a number of different contexts. The risk exposure in the context of asymmetric bimodal right-skewed precipitation data was further defined by using five essential risk indicators, such as value-at-risk, tail-value-at-risk, tail variance, tail mean–variance, and mean excess loss function. This was done in order to account for the right-skewed distribution of the data. In order to examine the data, several risk indicators were utilized. These risk indicators were used in order to achieve a more in-depth description of the risk exposure that was being faced.

本研究的目标是提出一种新的双参数寿命分布,并解释该分布的一些最基本特性。通过这项研究,我们将能够实现这两个目标。为了评估的目的,我们开展了利用模拟的研究,出于同样的原因,我们研究并考虑了各种不同的评估方法。利用两种不同的数据收集方式,可以分析建议的分配方式在不同情况下的适应性。在非对称双峰右斜降水数据的背景下,通过使用五个基本风险指标,如风险值、尾部风险值、尾部方差、尾部平均方差和平均超额损失函数,进一步定义了风险暴露。这样做是为了考虑数据的右偏分布。为了检验数据,使用了几个风险指标。使用这些风险指标是为了更深入地描述所面临的风险。
{"title":"A New Lindley Extension: Estimation, Risk Assessment and Analysis Under Bimodal Right Skewed Precipitation Data","authors":"Majid Hashempour,&nbsp;Morad Alizadeh,&nbsp;Haitham M. Yousof","doi":"10.1007/s40745-023-00485-1","DOIUrl":"10.1007/s40745-023-00485-1","url":null,"abstract":"<div><p>The objectives of this study are to propose a new two-parameter lifespan distribution and explain some of the most essential properties of that distribution. Through the course of this investigation, we will be able to achieve both of these objectives. For the aim of assessment, research is carried out that makes use of simulation, and for the same reason, a variety of various approaches are studied and taken into account for the purpose of evaluation. Making use of two separate data collections enables an analysis of the adaptability of the suggested distribution to a number of different contexts. The risk exposure in the context of asymmetric bimodal right-skewed precipitation data was further defined by using five essential risk indicators, such as value-at-risk, tail-value-at-risk, tail variance, tail mean–variance, and mean excess loss function. This was done in order to account for the right-skewed distribution of the data. In order to examine the data, several risk indicators were utilized. These risk indicators were used in order to achieve a more in-depth description of the risk exposure that was being faced.</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 6","pages":"1919 - 1958"},"PeriodicalIF":0.0,"publicationDate":"2023-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"42711550","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bayesian Analysis of Change Point Problems Using Conditionally Specified Priors 基于条件指定先验的变点问题的贝叶斯分析
Q1 Decision Sciences Pub Date : 2023-08-08 DOI: 10.1007/s40745-023-00484-2
G. Shahtahmassebi, José María Sarabia

In data analysis, change point problems correspond to abrupt changes in stochastic mechanisms generating data. The detection of change points is a relevant problem in the analysis and prediction of time series. In this paper, we consider a class of conjugate prior distributions obtained from conditional specification methodology for solving this problem. We illustrate the application of such distributions in Bayesian change point detection analysis with Poisson processes. We obtain the posterior distribution of model parameters using general bivariate distribution with gamma conditionals. Simulation from the posterior are readily implemented using a Gibbs sampling algorithm. The Gibbs sampling is implemented even when using conditional densities that are incompatible or only compatible with an improper joint density. The application of such methods will be demonstrated using examples of simulated and real data.

在数据分析中,变化点问题与产生数据的随机机制的突然变化相对应。变化点的检测是时间序列分析和预测中的一个相关问题。在本文中,我们考虑从条件规范方法中获得的一类共轭先验分布来解决这一问题。我们举例说明了这类分布在泊松过程的贝叶斯变化点检测分析中的应用。我们使用具有伽马条件的一般双变量分布来获得模型参数的后验分布。使用吉布斯采样算法可以很容易地从后验分布进行模拟。即使使用不相容或仅与不适当的联合密度相容的条件密度,也能实现吉布斯采样。我们将通过模拟数据和真实数据的例子来演示这些方法的应用。
{"title":"Bayesian Analysis of Change Point Problems Using Conditionally Specified Priors","authors":"G. Shahtahmassebi,&nbsp;José María Sarabia","doi":"10.1007/s40745-023-00484-2","DOIUrl":"10.1007/s40745-023-00484-2","url":null,"abstract":"<div><p>In data analysis, change point problems correspond to abrupt changes in stochastic mechanisms generating data. The detection of change points is a relevant problem in the analysis and prediction of time series. In this paper, we consider a class of conjugate prior distributions obtained from conditional specification methodology for solving this problem. We illustrate the application of such distributions in Bayesian change point detection analysis with Poisson processes. We obtain the posterior distribution of model parameters using general bivariate distribution with gamma conditionals. Simulation from the posterior are readily implemented using a Gibbs sampling algorithm. The Gibbs sampling is implemented even when using conditional densities that are incompatible or only compatible with an improper joint density. The application of such methods will be demonstrated using examples of simulated and real data.</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 6","pages":"1899 - 1918"},"PeriodicalIF":0.0,"publicationDate":"2023-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s40745-023-00484-2.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"47402928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bayesian Learning of Personalized Longitudinal Biomarker Trajectory 个性化纵向生物标志物轨迹的贝叶斯学习
Q1 Decision Sciences Pub Date : 2023-08-01 DOI: 10.1007/s40745-023-00486-0
Shouhao Zhou, Xuelin Huang, Chan Shen, Hagop M. Kantarjian

This work concerns the effective personalized prediction of longitudinal biomarker trajectory, motivated by a study of cancer targeted therapy for patients with chronic myeloid leukemia (CML). Continuous monitoring with a confirmed biomarker of residual disease is a key component of CML management for early prediction of disease relapse. However, the longitudinal biomarker measurements have highly heterogeneous trajectories between subjects (patients) with various shapes and patterns. It is believed that the trajectory is clinically related to the development of treatment resistance, but there was limited knowledge about the underlying mechanism. To address the challenge, we propose a novel Bayesian approach to modeling the distribution of subject-specific longitudinal trajectories. It exploits flexible Bayesian learning to accommodate complex changing patterns over time and non-linear covariate effects, and allows for real-time prediction of both in-sample and out-of-sample subjects. The generated information can help make clinical decisions, and consequently enhance the personalized treatment management of precision medicine.

这项工作涉及对纵向生物标志物轨迹进行有效的个性化预测,其动机是对慢性髓性白血病(CML)患者进行癌症靶向治疗研究。用确诊的生物标志物对残留疾病进行持续监测是 CML 治疗的关键组成部分,可用于疾病复发的早期预测。然而,不同受试者(患者)之间的纵向生物标志物测量结果具有高度异质性的轨迹,形状和模式各不相同。人们认为这种轨迹在临床上与治疗耐药性的发展有关,但对其潜在机制的了解却很有限。为了应对这一挑战,我们提出了一种新颖的贝叶斯方法来模拟受试者特定纵向轨迹的分布。它利用灵活的贝叶斯学习来适应复杂的随时间变化的模式和非线性协变量效应,并允许对样本内和样本外受试者进行实时预测。生成的信息有助于临床决策,从而加强精准医学的个性化治疗管理。
{"title":"Bayesian Learning of Personalized Longitudinal Biomarker Trajectory","authors":"Shouhao Zhou,&nbsp;Xuelin Huang,&nbsp;Chan Shen,&nbsp;Hagop M. Kantarjian","doi":"10.1007/s40745-023-00486-0","DOIUrl":"10.1007/s40745-023-00486-0","url":null,"abstract":"<div><p>This work concerns the effective personalized prediction of longitudinal biomarker trajectory, motivated by a study of cancer targeted therapy for patients with chronic myeloid leukemia (CML). Continuous monitoring with a confirmed biomarker of residual disease is a key component of CML management for early prediction of disease relapse. However, the longitudinal biomarker measurements have highly heterogeneous trajectories between subjects (patients) with various shapes and patterns. It is believed that the trajectory is clinically related to the development of treatment resistance, but there was limited knowledge about the underlying mechanism. To address the challenge, we propose a novel Bayesian approach to modeling the distribution of subject-specific longitudinal trajectories. It exploits flexible Bayesian learning to accommodate complex changing patterns over time and non-linear covariate effects, and allows for real-time prediction of both in-sample and out-of-sample subjects. The generated information can help make clinical decisions, and consequently enhance the personalized treatment management of precision medicine.</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 3","pages":"1031 - 1050"},"PeriodicalIF":0.0,"publicationDate":"2023-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"46463104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Applications of Reliability Test Plan for Logistic Rayleigh Distributed Quality Characteristic 物流瑞利分布质量特性可靠性试验计划的应用
Q1 Decision Sciences Pub Date : 2023-07-19 DOI: 10.1007/s40745-023-00473-5
Mahendra Saha, Harsh Tripathi, Anju Devi, Pratibha Pareek

In this article, a reliability test plan under time truncated life test is considered for the logistic Rayleigh distribution ((mathcal {LRD})). A brief discussion over statistical properties and significance of the (mathcal {LRD}) is placed in this present study. Larger the value of median—better is the quality of the lot is considered as quality characteristic for the proposed reliability test plan. Minimum sample sizes are placed in tabular form for different set up of specified consumer’s risk. Also operating characteristics ((mathcal{O}mathcal{C})) values are shown in tabular forms for the chosen set up and discussed the pattern of (mathcal{O}mathcal{C}) values. A comparative analysis of the present study with some other reliability test plans is discussed based on the sample sizes. As an illustration, the performance of the proposed plan for the (mathcal {LRD}) is shown through real-life examples.

本文考虑了时间截断寿命试验下的可靠性试验计划,即 logistic Rayleigh 分布((mathcal {LRD}))。本文简要讨论了 (mathcal {LRD}) 的统计特性和意义。中值越大,批次质量越好,这被认为是建议的可靠性测试计划的质量特征。针对不同的消费者风险设置,最小样本量以表格形式列出。此外,还以表格形式显示了所选设置的运行特征((mathcal{O}mathcal{C}))值,并讨论了(mathcal{O}mathcal{C})值的模式。根据样本量,讨论了本研究与其他一些可靠性测试计划的比较分析。作为说明,通过实际例子展示了所建议的计划在 (mathcal {LRD}) 方面的性能。
{"title":"Applications of Reliability Test Plan for Logistic Rayleigh Distributed Quality Characteristic","authors":"Mahendra Saha,&nbsp;Harsh Tripathi,&nbsp;Anju Devi,&nbsp;Pratibha Pareek","doi":"10.1007/s40745-023-00473-5","DOIUrl":"10.1007/s40745-023-00473-5","url":null,"abstract":"<div><p>In this article, a reliability test plan under time truncated life test is considered for the logistic Rayleigh distribution (<span>(mathcal {LRD})</span>). A brief discussion over statistical properties and significance of the <span>(mathcal {LRD})</span> is placed in this present study. Larger the value of median—better is the quality of the lot is considered as quality characteristic for the proposed reliability test plan. Minimum sample sizes are placed in tabular form for different set up of specified consumer’s risk. Also operating characteristics (<span>(mathcal{O}mathcal{C})</span>) values are shown in tabular forms for the chosen set up and discussed the pattern of <span>(mathcal{O}mathcal{C})</span> values. A comparative analysis of the present study with some other reliability test plans is discussed based on the sample sizes. As an illustration, the performance of the proposed plan for the <span>(mathcal {LRD})</span> is shown through real-life examples.</p></div>","PeriodicalId":36280,"journal":{"name":"Annals of Data Science","volume":"11 5","pages":"1687 - 1703"},"PeriodicalIF":0.0,"publicationDate":"2023-07-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"43684187","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Annals of Data Science
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1