Pub Date : 2026-02-01Epub Date: 2026-02-12DOI: 10.1177/2167647X261423114
Hwijae Son, Sung Woong Cho, Hyung Ju Hwang
Recommender systems play a crucial role in enhancing user engagement across domains such as e-commerce, social media, entertainment, and education. Recently, they have also been used in marketing to identify high-value customers and personalize campaigns. However, small businesses often struggle with the high cost per action (CPA) and low conversion rates (CR) associated with online marketing platforms. To address this challenge, we propose a novel recommender system that leverages offline interaction data to identify customers likely to use discount coupons, thereby increasing CRs and reducing marketing costs. We address technical challenges such as cold-start problems and data sparsity by introducing tailored data augmentation techniques. The effectiveness of our approach is validated through experiments using store-level coupon and point log data, evaluated with metrics including CPA, CR, and root mean squared error. Results show that our system significantly outperforms conventional online marketing platforms, emphasizing the value of incorporating offline data with proper augmentation for cost-effective marketing.
{"title":"Offline Data-Driven Recommender Systems for Improving Small Business Marketing Strategies.","authors":"Hwijae Son, Sung Woong Cho, Hyung Ju Hwang","doi":"10.1177/2167647X261423114","DOIUrl":"10.1177/2167647X261423114","url":null,"abstract":"<p><p>Recommender systems play a crucial role in enhancing user engagement across domains such as e-commerce, social media, entertainment, and education. Recently, they have also been used in marketing to identify high-value customers and personalize campaigns. However, small businesses often struggle with the high cost per action (CPA) and low conversion rates (CR) associated with online marketing platforms. To address this challenge, we propose a novel recommender system that leverages offline interaction data to identify customers likely to use discount coupons, thereby increasing CRs and reducing marketing costs. We address technical challenges such as cold-start problems and data sparsity by introducing tailored data augmentation techniques. The effectiveness of our approach is validated through experiments using store-level coupon and point log data, evaluated with metrics including CPA, CR, and root mean squared error. Results show that our system significantly outperforms conventional online marketing platforms, emphasizing the value of incorporating offline data with proper augmentation for cost-effective marketing.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":" ","pages":"1-12"},"PeriodicalIF":2.6,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146168160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2026-02-01Epub Date: 2025-12-19DOI: 10.1177/2167647X251399169
Qiong He, Zhenwei Yang, Yijia Li
Enhancing brand value is critical for new energy vehicle (NEV) enterprises amid fierce competition. This study leverages online consumer reviews as core big data to drive brand equity improvement via advanced big data analytics. A large-scale dataset of 5564 reviews for top five best-selling NEVs was collected from "Dongche Di" via web scraping, followed by a big data processing pipeline (data cleaning, Jieba segmentation, and stop-word filtering). To mine unstructured text big data, we used word cloud visualization, semantic network analysis, and an Latent Dirichlet Allocation (LDA)-Long Short-Term Memory (LSTM) fusion model: LDA identified key consumer concern dimensions, while LSTM enabled deep sentiment classification. Big data analysis revealed five core NEV brand perception dimensions (range, driving experience, interior space, price, and high-speed performance) and quantified emotions-prominent negativity in driving experience, minimal negativity in interior space, and overall dominant negativity. Guided by the Consumer-Based Brand Equity model, we proposed brand enhancement strategies. This study showcases big data analytics' power in scaling consumer perception understanding, offering a data-centric framework for NEV firms to optimize branding.
{"title":"Enhancing NEV Brand Equity Through Big Data Analytics: An LDA-LSTM Approach to Mining Online Consumer Reviews.","authors":"Qiong He, Zhenwei Yang, Yijia Li","doi":"10.1177/2167647X251399169","DOIUrl":"10.1177/2167647X251399169","url":null,"abstract":"<p><p>Enhancing brand value is critical for new energy vehicle (NEV) enterprises amid fierce competition. This study leverages online consumer reviews as core big data to drive brand equity improvement via advanced big data analytics. A large-scale dataset of 5564 reviews for top five best-selling NEVs was collected from \"Dongche Di\" via web scraping, followed by a big data processing pipeline (data cleaning, Jieba segmentation, and stop-word filtering). To mine unstructured text big data, we used word cloud visualization, semantic network analysis, and an Latent Dirichlet Allocation (LDA)-Long Short-Term Memory (LSTM) fusion model: LDA identified key consumer concern dimensions, while LSTM enabled deep sentiment classification. Big data analysis revealed five core NEV brand perception dimensions (range, driving experience, interior space, price, and high-speed performance) and quantified emotions-prominent negativity in driving experience, minimal negativity in interior space, and overall dominant negativity. Guided by the Consumer-Based Brand Equity model, we proposed brand enhancement strategies. This study showcases big data analytics' power in scaling consumer perception understanding, offering a data-centric framework for NEV firms to optimize branding.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":" ","pages":"42-55"},"PeriodicalIF":2.6,"publicationDate":"2026-02-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145858991","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-12DOI: 10.1177/2167647X251403895
Zhaodi Yu, Zhenxiang Xu, Jiangang Qi
In the context of a global risk society, emergency law has become a critical field for balancing the expansion of state power with the protection of civil rights during crises. Despite its growing importance, a systematic, quantitative comparison of the knowledge landscapes of international and Chinese emergency law scholarship has been notably absent. This study employs bibliometric and knowledge mapping analysis, utilizing CiteSpace software. A total of 274 publications were retrieved from the Web of Science Core Collection and 391 from the China National Knowledge Infrastructure database. These data were used to systematically map and compare the research status, collaborative networks, and core themes of the two academic communities. The findings indicate that while both international and Chinese research are crisis-driven, with publication surges corresponding to major events such as the 9/11 attacks, SARS, and the COVID-19 pandemic, they function as two academically isolated communities with no author-level collaboration. A fundamental divergence in research paradigms was identified. International scholarship follows a "limitation-oriented" paradigm, rooted in liberal constitutionalism, focusing on the tension between emergency powers and human rights, and the risks of a state of exception. In contrast, Chinese research adopts a "construction-oriented" paradigm aimed at building an efficient, state-centric crisis response system, dominated by concepts such as emergency management and the "one plan and three sub-systems" framework. This study concludes that there are two worlds of emergency law. The international paradigm primarily treats emergency law as a mechanism to constrain state authority and protect individual rights from government overreach. In contrast, the Chinese paradigm views law as an instrument to enhance state capacity and ensure effective crisis management. This fundamental divergence in normative goals and theoretical foundations identified in this study presents significant theoretical and practical challenges for global emergency governance and offers a clear direction for future comparative legal studies.
在全球风险社会背景下,紧急状态法已成为在危机中平衡国家权力扩张与公民权利保护的关键领域。尽管其重要性日益增加,但对国际和中国紧急法学术知识格局的系统、定量比较明显缺乏。本研究采用文献计量学和知识图谱分析法,利用CiteSpace软件。共检索到Web of Science核心文献274篇,检索到中国国家知识基础设施数据库391篇。这些数据被用于系统地绘制和比较两个学术界的研究现状、合作网络和核心主题。研究结果表明,虽然国际和中国的研究都是危机驱动的,发表量激增对应于9/11袭击、SARS和COVID-19大流行等重大事件,但它们在学术上是两个孤立的社区,没有作者层面的合作。研究范式存在根本性分歧。国际学术遵循一种“以限制为导向”的范式,根植于自由宪政主义,关注紧急权力与人权之间的紧张关系,以及例外状态的风险。相比之下,中国的研究采用“建构导向”的范式,旨在构建一个高效的、以国家为中心的危机应对体系,以应急管理和“一计划三子系统”框架等概念为主导。本研究的结论是,紧急状态法有两个世界。国际范例主要将紧急状态法视为一种约束国家权威和保护个人权利免受政府越权的机制。相比之下,中国范式将法律视为提高国家能力和确保有效危机管理的工具。本研究确定的规范目标和理论基础的根本分歧为全球应急治理提出了重大的理论和实践挑战,并为未来的比较法律研究提供了明确的方向。
{"title":"The Two Worlds of Emergency Law: A Comparative Study of International and Chinese Scholarship Through Knowledge Domain Mapping.","authors":"Zhaodi Yu, Zhenxiang Xu, Jiangang Qi","doi":"10.1177/2167647X251403895","DOIUrl":"https://doi.org/10.1177/2167647X251403895","url":null,"abstract":"<p><p>In the context of a global risk society, emergency law has become a critical field for balancing the expansion of state power with the protection of civil rights during crises. Despite its growing importance, a systematic, quantitative comparison of the knowledge landscapes of international and Chinese emergency law scholarship has been notably absent. This study employs bibliometric and knowledge mapping analysis, utilizing CiteSpace software. A total of 274 publications were retrieved from the Web of Science Core Collection and 391 from the China National Knowledge Infrastructure database. These data were used to systematically map and compare the research status, collaborative networks, and core themes of the two academic communities. The findings indicate that while both international and Chinese research are crisis-driven, with publication surges corresponding to major events such as the 9/11 attacks, SARS, and the COVID-19 pandemic, they function as two academically isolated communities with no author-level collaboration. A fundamental divergence in research paradigms was identified. International scholarship follows a \"limitation-oriented\" paradigm, rooted in liberal constitutionalism, focusing on the tension between emergency powers and human rights, and the risks of a state of exception. In contrast, Chinese research adopts a \"construction-oriented\" paradigm aimed at building an efficient, state-centric crisis response system, dominated by concepts such as emergency management and the \"one plan and three sub-systems\" framework. This study concludes that there are two worlds of emergency law. The international paradigm primarily treats emergency law as a mechanism to constrain state authority and protect individual rights from government overreach. In contrast, the Chinese paradigm views law as an instrument to enhance state capacity and ensure effective crisis management. This fundamental divergence in normative goals and theoretical foundations identified in this study presents significant theoretical and practical challenges for global emergency governance and offers a clear direction for future comparative legal studies.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":" ","pages":""},"PeriodicalIF":2.6,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145835330","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-08-22DOI: 10.1177/2167647X251366060
Xuna Wang
With the rapid development of social media and online platforms, the speed and influence of emergency dissemination in cyberspace have significantly increased. The swift changes in public opinion, especially the phenomenon of opinion reversals, exert profound impacts on social stability and government credibility. The hypernetwork structure, characterized by its multilayered and multidimensional complexity, offers a new theoretical framework for analyzing multiagents and their interactions in the evolution of public opinion. Based on hypernetwork theory, this study constructs a four-layer subnet model encompassing user interaction network, event evolution network, semantic association network, and emotional conduction network. By extracting network structural features and conducting cross-layer linkage analysis, an identification system for public opinion reversals in emergencies is established. Taking the donation incident involving Hongxing Erke during the Henan rainstorm in 2021 as a case study, an empirical analysis of the public opinion reversal process is conducted. The research results indicate that the proposed hypernetwork model can effectively identify key nodes in public opinion reversals. The multi-indicator collaborative identification system for public opinion reversals aids in rapidly and effectively detecting signals of such reversals. This study not only provides new methodological support for the dynamic identification of public opinion reversals but also offers theoretical references and practical guidance for public opinion monitoring and emergency response decision-making in emergencies.
{"title":"A Study of Public Opinion Reversal Recognition of Emergency Based on Hypernetwork.","authors":"Xuna Wang","doi":"10.1177/2167647X251366060","DOIUrl":"10.1177/2167647X251366060","url":null,"abstract":"<p><p>With the rapid development of social media and online platforms, the speed and influence of emergency dissemination in cyberspace have significantly increased. The swift changes in public opinion, especially the phenomenon of opinion reversals, exert profound impacts on social stability and government credibility. The hypernetwork structure, characterized by its multilayered and multidimensional complexity, offers a new theoretical framework for analyzing multiagents and their interactions in the evolution of public opinion. Based on hypernetwork theory, this study constructs a four-layer subnet model encompassing user interaction network, event evolution network, semantic association network, and emotional conduction network. By extracting network structural features and conducting cross-layer linkage analysis, an identification system for public opinion reversals in emergencies is established. Taking the donation incident involving Hongxing Erke during the Henan rainstorm in 2021 as a case study, an empirical analysis of the public opinion reversal process is conducted. The research results indicate that the proposed hypernetwork model can effectively identify key nodes in public opinion reversals. The multi-indicator collaborative identification system for public opinion reversals aids in rapidly and effectively detecting signals of such reversals. This study not only provides new methodological support for the dynamic identification of public opinion reversals but also offers theoretical references and practical guidance for public opinion monitoring and emergency response decision-making in emergencies.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":" ","pages":"497-512"},"PeriodicalIF":2.6,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144977778","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01DOI: 10.1177/2167647X251406607
Yuping Yan, Hanyang Xie, Liang Chen, You Wen, Huaquan Su
Data in power grid digital operation exhibit multisource heterogeneous characteristics, resulting in low integration efficiency and slow anomaly detection response. To address this, this paper proposes a method for power grid digital operation data integration based on K-medoids clustering. The basic service layer utilizes an Field Programmable Gate Array parallel architecture. This enables millisecond-level synchronous acquisition and dynamic preprocessing of multisource data, such as mechanical vibration, partial discharge signals, and temperature. The implementation is based on the analysis of the power grid digital operation structure. The data are then fed back to the cloud service layer, which, through business integration services, data analysis, and data access services, performs data filtering and analysis. Subsequently, the data are input to the application layer via the database server. The application layer employs a K-medoids clustering method that introduces a density-weighted Euclidean distance metric and an adaptive centroid selection strategy, significantly enhancing the clustering performance of multisource data. In particular, the proposed architecture supports real-time data processing and can be extended to cross-modal scenarios, including integration with speech-to-text systems in power grid monitoring. By aligning with low-latency neural network principles, this method facilitates timely decision-making in intelligent operation environments. Experiments confirm the method's efficacy. It acquires and integrates multisource heterogeneous power grid digital operation data effectively. The data throughput of different power grid digital operation data sources all exceed 110 MB/s. The silhouette coefficient of the integrated data sets is greater than 0.91, indicating that the integration of power grid digital operation data using this method exhibits good separability and reliability, enabling rapid detection of data anomalies within the power grid, thus laying a solid foundation for the operation and maintenance management of power grid digital operation.
{"title":"Method for Power Grid Digital Operation Data Integration Based on K-Medoids Clustering with Support for Real-Time Cross-Modal Applications.","authors":"Yuping Yan, Hanyang Xie, Liang Chen, You Wen, Huaquan Su","doi":"10.1177/2167647X251406607","DOIUrl":"https://doi.org/10.1177/2167647X251406607","url":null,"abstract":"<p><p>Data in power grid digital operation exhibit multisource heterogeneous characteristics, resulting in low integration efficiency and slow anomaly detection response. To address this, this paper proposes a method for power grid digital operation data integration based on K-medoids clustering. The basic service layer utilizes an Field Programmable Gate Array parallel architecture. This enables millisecond-level synchronous acquisition and dynamic preprocessing of multisource data, such as mechanical vibration, partial discharge signals, and temperature. The implementation is based on the analysis of the power grid digital operation structure. The data are then fed back to the cloud service layer, which, through business integration services, data analysis, and data access services, performs data filtering and analysis. Subsequently, the data are input to the application layer via the database server. The application layer employs a K-medoids clustering method that introduces a density-weighted Euclidean distance metric and an adaptive centroid selection strategy, significantly enhancing the clustering performance of multisource data. In particular, the proposed architecture supports real-time data processing and can be extended to cross-modal scenarios, including integration with speech-to-text systems in power grid monitoring. By aligning with low-latency neural network principles, this method facilitates timely decision-making in intelligent operation environments. Experiments confirm the method's efficacy. It acquires and integrates multisource heterogeneous power grid digital operation data effectively. The data throughput of different power grid digital operation data sources all exceed 110 MB/s. The silhouette coefficient of the integrated data sets is greater than 0.91, indicating that the integration of power grid digital operation data using this method exhibits good separability and reliability, enabling rapid detection of data anomalies within the power grid, thus laying a solid foundation for the operation and maintenance management of power grid digital operation.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":"13 6","pages":"453-470"},"PeriodicalIF":2.6,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145716573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Soybeans are a high-quality vegetable protein resource and a fundamental strategic material integral to the national economy and public livelihood. To investigate the research status of soybean quality evaluation, this study analyzes relevant literature from Web of Science and China Knowledge Network (2000-2024). Using bibliometric methods with Excel and VOSviewer, we examined publication years, keywords, authors, sources, countries/regions, and institutions, generating visualizations to intuitively illustrate the field's developmental status. Results indicate that over the past 25 years, soybean quality evaluation research has emerged as a focal point in crop science, with institutions predominantly located in China and the United States. Key journals in this domain include Food Chemistry, Frontiers in Plant Science, and Soybean Science, among others. Research primarily focuses on soybean physical characteristics and the component-quality relationship. Interdisciplinary advancements have positioned spectral analysis, intelligent systems, and multitechnology fusion as innovative frontiers in this field. These findings enhance researchers' understanding of current trends and support evidence-based decision-making in soybean quality evaluation.
大豆是一种优质植物蛋白资源,是关系国计民生的基础性战略物资。为了了解大豆品质评价的研究现状,本研究分析了Web of Science和中国知识网(2000-2024)的相关文献。利用文献计量学方法,结合Excel和VOSviewer,对论文的出版年份、关键词、作者、来源、国家/地区和机构进行了统计分析,生成了可视化图,直观地说明了该领域的发展状况。结果表明,在过去的25年中,大豆质量评价研究已成为作物科学的一个焦点,研究机构主要集中在中国和美国。该领域的主要期刊包括《食品化学》、《植物科学前沿》和《大豆科学》等。研究主要集中在大豆的物理特性和成分与品质的关系。跨学科的进步将光谱分析、智能系统和多技术融合定位为该领域的创新前沿。这些发现增强了研究人员对当前趋势的理解,并为大豆质量评价的循证决策提供了支持。
{"title":"Analysis on Research Situation of Soybean Quality Evaluation Based on Bibliometrics.","authors":"Yanxia Gao, Pengju Tang, Xuhong Tang, Dong Wang, Jiaqi Luo, JiaDong Wu","doi":"10.1177/2167647X251399053","DOIUrl":"10.1177/2167647X251399053","url":null,"abstract":"<p><p>Soybeans are a high-quality vegetable protein resource and a fundamental strategic material integral to the national economy and public livelihood. To investigate the research status of soybean quality evaluation, this study analyzes relevant literature from Web of Science and China Knowledge Network (2000-2024). Using bibliometric methods with Excel and VOSviewer, we examined publication years, keywords, authors, sources, countries/regions, and institutions, generating visualizations to intuitively illustrate the field's developmental status. Results indicate that over the past 25 years, soybean quality evaluation research has emerged as a focal point in crop science, with institutions predominantly located in China and the United States. Key journals in this domain include <i>Food Chemistry</i>, <i>Frontiers in Plant Science</i>, and <i>Soybean Science</i>, among others. Research primarily focuses on soybean physical characteristics and the component-quality relationship. Interdisciplinary advancements have positioned spectral analysis, intelligent systems, and multitechnology fusion as innovative frontiers in this field. These findings enhance researchers' understanding of current trends and support evidence-based decision-making in soybean quality evaluation.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":" ","pages":"487-496"},"PeriodicalIF":2.6,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145679308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01DOI: 10.1177/2167647X251405797
Qiong He, Xueqing Guo
This study aims to enhance the prediction precision of aircraft engine remaining useful life (RUL) by overcoming common challenges in current models, such as ineffective feature extraction and insufficient modeling of long-term temporal dependencies. We propose a novel multilayer hybrid architecture that combines bidirectional long short-term memory (BiLSTM) and gated recurrent unit (GRU) networks, augmented with an attention mechanism to enhance the model's focus on informative temporal patterns. In this framework, raw time series data are initially processed by the BiLSTM to extract bidirectional features associated with engine health conditions. The GRU network is subsequently used to effectively model long-range dependencies, thereby enriching the temporal representation. An adaptive attention module is included to assign varying importance to different features, allowing the model to focus on key indicators of engine condition. Evaluation results on the FD001 and FD003 datasets show that the model achieves root mean squared error reductions ranging from 8.81% to 30.60% and from 7.48% to 37.96%, validating its performance and robustness in RUL forecasting. In comparison with conventional BiLSTM and GRU models, the proposed BiLSTM-GRU-Attention architecture integrates attention-based feature weighting with a hybrid recurrent framework, thereby offering a concise and effective approach to RUL prediction for aircraft engines.
{"title":"Prediction of Remaining Life of Aircraft Engines Based on BiLSTM-GRU-Attention Model.","authors":"Qiong He, Xueqing Guo","doi":"10.1177/2167647X251405797","DOIUrl":"https://doi.org/10.1177/2167647X251405797","url":null,"abstract":"<p><p>This study aims to enhance the prediction precision of aircraft engine remaining useful life (RUL) by overcoming common challenges in current models, such as ineffective feature extraction and insufficient modeling of long-term temporal dependencies. We propose a novel multilayer hybrid architecture that combines bidirectional long short-term memory (BiLSTM) and gated recurrent unit (GRU) networks, augmented with an attention mechanism to enhance the model's focus on informative temporal patterns. In this framework, raw time series data are initially processed by the BiLSTM to extract bidirectional features associated with engine health conditions. The GRU network is subsequently used to effectively model long-range dependencies, thereby enriching the temporal representation. An adaptive attention module is included to assign varying importance to different features, allowing the model to focus on key indicators of engine condition. Evaluation results on the FD001 and FD003 datasets show that the model achieves root mean squared error reductions ranging from 8.81% to 30.60% and from 7.48% to 37.96%, validating its performance and robustness in RUL forecasting. In comparison with conventional BiLSTM and GRU models, the proposed BiLSTM-GRU-Attention architecture integrates attention-based feature weighting with a hybrid recurrent framework, thereby offering a concise and effective approach to RUL prediction for aircraft engines.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":"13 6","pages":"471-486"},"PeriodicalIF":2.6,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145716512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-01Epub Date: 2025-11-18DOI: 10.1177/2167647X251392796
Yang Wang, Tianchun Xiang, Shuai Luo, Yi Gao, Xiangyu Kong
Human activities that generate greenhouse gas emissions pose a significant threat to urban green and sustainable development. Production activities in key industrial sectors are a primary contributor to high urban carbon emissions. Therefore, effectively reducing carbon emissions in these sectors is crucial for achieving urban carbon peak and neutrality goals. Carbon emission monitoring is a critical approach that aids governmental bodies in understanding changes in industrial carbon emissions, thereby supporting decision-making and carbon reduction efforts. However, current industry-oriented carbon monitoring methods suffer from issues such as low frequency, poor accuracy, and inadequate privacy security. To address these challenges, this article proposes a novel privacy-protected "electricity-carbon'' nexus model, long short-term memory with the vertical federated framework (VF-LSTM), to monitor carbon emissions in key urban industries. The vertical federated framework ensures "usable but invisible" privacy protection for multisource data from various participants. The embedded long short-term memory model accurately captures industry-specific carbon emissions. Using data from key industries (steel, petrochemical, chemical, and nonferrous industries), this article constructs and validates the performance of the proposed industry-level carbon emission monitoring model. The results demonstrate that the model has high accuracy and robustness, effectively monitoring industry carbon emissions while protecting data privacy.
{"title":"Monitoring Carbon Emission from Key Industries Based on VF-LSTM Model.","authors":"Yang Wang, Tianchun Xiang, Shuai Luo, Yi Gao, Xiangyu Kong","doi":"10.1177/2167647X251392796","DOIUrl":"10.1177/2167647X251392796","url":null,"abstract":"<p><p>Human activities that generate greenhouse gas emissions pose a significant threat to urban green and sustainable development. Production activities in key industrial sectors are a primary contributor to high urban carbon emissions. Therefore, effectively reducing carbon emissions in these sectors is crucial for achieving urban carbon peak and neutrality goals. Carbon emission monitoring is a critical approach that aids governmental bodies in understanding changes in industrial carbon emissions, thereby supporting decision-making and carbon reduction efforts. However, current industry-oriented carbon monitoring methods suffer from issues such as low frequency, poor accuracy, and inadequate privacy security. To address these challenges, this article proposes a novel privacy-protected \"electricity-carbon'' nexus model, long short-term memory with the vertical federated framework (VF-LSTM), to monitor carbon emissions in key urban industries. The vertical federated framework ensures \"usable but invisible\" privacy protection for multisource data from various participants. The embedded long short-term memory model accurately captures industry-specific carbon emissions. Using data from key industries (steel, petrochemical, chemical, and nonferrous industries), this article constructs and validates the performance of the proposed industry-level carbon emission monitoring model. The results demonstrate that the model has high accuracy and robustness, effectively monitoring industry carbon emissions while protecting data privacy.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":" ","pages":"441-452"},"PeriodicalIF":2.6,"publicationDate":"2025-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145574953","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-02-28DOI: 10.1089/big.2024.0128
Ikpe Justice Akpan, Rouzbeh Razavi, Asuama A Akpan
Decision sciences (DSC) involves studying complex dynamic systems and processes to aid informed choices subject to constraints in uncertain conditions. It integrates multidisciplinary methods and strategies to evaluate decision engineering processes, identifying alternatives and providing insights toward enhancing prudent decision-making. This study analyzes the evolutionary trends and innovation in DSC education and research trends over the past 25 years. Using metadata from bibliographic records and employing the science mapping method and text analytics, we map and evaluate the thematic, intellectual, and social structures of DSC research. The results identify "knowledge management," "decision support systems," "data envelopment analysis," "simulation," and "artificial intelligence" (AI) as some of the prominent critical skills and knowledge requirements for problem-solving in DSC before and during the period (2000-2024). However, these technologies are evolving significantly in the recent wave of digital transformation, with data analytics frameworks (including techniques such as big data analytics, machine learning, business intelligence, data mining, and information visualization) becoming crucial. DSC education and research continue to mirror the development in practice, with sustainable education through virtual/online learning becoming prominent. Innovative pedagogical approaches/strategies also include computer simulation and games ("play and learn" or "role-playing"). The current era witnesses AI adoption in different forms as conversational Chatbot agent and generative AI (GenAI), such as chat generative pretrained transformer in teaching, learning, and scholarly activities amidst challenges (academic integrity, plagiarism, intellectual property violations, and other ethical and legal issues). Future DSC education must innovatively integrate GenAI into DSC education and address the resulting challenges.
{"title":"Evolutionary Trends in Decision Sciences Education Research from Simulation and Games to Big Data Analytics and Generative Artificial Intelligence.","authors":"Ikpe Justice Akpan, Rouzbeh Razavi, Asuama A Akpan","doi":"10.1089/big.2024.0128","DOIUrl":"10.1089/big.2024.0128","url":null,"abstract":"<p><p>Decision sciences (DSC) involves studying complex dynamic systems and processes to aid informed choices subject to constraints in uncertain conditions. It integrates multidisciplinary methods and strategies to evaluate decision engineering processes, identifying alternatives and providing insights toward enhancing prudent decision-making. This study analyzes the evolutionary trends and innovation in DSC education and research trends over the past 25 years. Using metadata from bibliographic records and employing the science mapping method and text analytics, we map and evaluate the thematic, intellectual, and social structures of DSC research. The results identify \"knowledge management,\" \"decision support systems,\" \"data envelopment analysis,\" \"simulation,\" and \"artificial intelligence\" (AI) as some of the prominent critical skills and knowledge requirements for problem-solving in DSC before and during the period (2000-2024). However, these technologies are evolving significantly in the recent wave of digital transformation, with data analytics frameworks (including techniques such as big data analytics, machine learning, business intelligence, data mining, and information visualization) becoming crucial. DSC education and research continue to mirror the development in practice, with sustainable education through virtual/online learning becoming prominent. Innovative pedagogical approaches/strategies also include computer simulation and games (\"play and learn\" or \"role-playing\"). The current era witnesses AI adoption in different forms as conversational Chatbot agent and generative AI (GenAI), such as chat generative pretrained transformer in teaching, learning, and scholarly activities amidst challenges (academic integrity, plagiarism, intellectual property violations, and other ethical and legal issues). Future DSC education must innovatively integrate GenAI into DSC education and address the resulting challenges.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":" ","pages":"416-437"},"PeriodicalIF":2.6,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143527974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-10-01Epub Date: 2025-01-10DOI: 10.1089/big.2024.0036
Sofie Goethals, Sandra Matz, Foster Provost, David Martens, Yanou Ramon
Our online lives generate a wealth of behavioral records-digital footprints-which are stored and leveraged by technology platforms. These data can be used to create value for users by personalizing services. At the same time, however, it also poses a threat to people's privacy by offering a highly intimate window into their private traits (e.g., their personality, political ideology, sexual orientation). We explore the concept of cloaking: allowing users to hide parts of their digital footprints from predictive algorithms, to prevent unwanted inferences. This article addresses two open questions: (i) can cloaking be effective in the longer term, as users continue to generate new digital footprints? And (ii) what is the potential impact of cloaking on the accuracy of desirable inferences? We introduce a novel strategy focused on cloaking "metafeatures" and compare its efficacy against just cloaking the raw footprints. The main findings are (i) while cloaking effectiveness does indeed diminish over time, using metafeatures slows the degradation; (ii) there is a tradeoff between privacy and personalization: cloaking undesired inferences also can inhibit desirable inferences. Furthermore, the metafeature strategy-which yields more stable cloaking-also incurs a larger reduction in desirable inferences.
{"title":"The Impact of Cloaking Digital Footprints on User Privacy and Personalization.","authors":"Sofie Goethals, Sandra Matz, Foster Provost, David Martens, Yanou Ramon","doi":"10.1089/big.2024.0036","DOIUrl":"10.1089/big.2024.0036","url":null,"abstract":"<p><p>Our online lives generate a wealth of behavioral records-<i>digital footprints</i>-which are stored and leveraged by technology platforms. These data can be used to create value for users by personalizing services. At the same time, however, it also poses a threat to people's privacy by offering a highly intimate window into their private traits (e.g., their personality, political ideology, sexual orientation). We explore the concept of <i>cloaking</i>: allowing users to hide parts of their digital footprints from predictive algorithms, to prevent unwanted inferences. This article addresses two open questions: (i) can cloaking be effective in the longer term, as users continue to generate new digital footprints? And (ii) what is the potential impact of cloaking on the accuracy of <i>desirable</i> inferences? We introduce a novel strategy focused on cloaking \"metafeatures\" and compare its efficacy against just cloaking the raw footprints. The main findings are (i) while cloaking effectiveness does indeed diminish over time, using metafeatures slows the degradation; (ii) there is a tradeoff between privacy and personalization: cloaking undesired inferences also can inhibit desirable inferences. Furthermore, the metafeature strategy-which yields more stable cloaking-also incurs a larger reduction in desirable inferences.</p>","PeriodicalId":51314,"journal":{"name":"Big Data","volume":" ","pages":"345-363"},"PeriodicalIF":2.6,"publicationDate":"2025-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142958560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}