首页 > 最新文献

IEEE open journal of signal processing最新文献

英文 中文
Uncertainty-Based Streaming ASR With Evidential Deep Learning 基于不确定性的流ASR与证据深度学习
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-23 DOI: 10.1109/OJSP.2026.3657308
Hiroaki Sato;Asahi Sakuma;Ryuga Sugano;Tadashi Kumano;Yoshihiko Kawai;Shinji Watanabe;Tetsuji Ogawa
Attention-based encoder-decoder (AED) models achieve high accuracy in offline automatic speech recognition (ASR), but their application to streaming remains challenging due to the lack of mechanisms for regulating token emission. Existing approaches include monotonic attention, forced alignment with external models providing token-level boundaries, and encoder-based emission control methods. However, these methods either require structural modifications, complicate the training pipeline, or show limited accuracy. In addition, local agreement has been proposed as a method enabling streaming without retraining, but it incurs fixed delays corresponding to the input window size and premature commitments. To address these limitations, we propose Evidential Streaming TRAnsformer (ESTRA), a framework that leverages evidential deep learning (EDL) to estimate uncertainty. ESTRA models token probabilities with a Dirichlet distribution and introduces hierarchical and direct Kullback–Leibler divergence losses to ensure uncertainty decreases progressively as more speech is observed. During inference, token emission is controlled by comparing uncertainty against a threshold, suppressing premature outputs without fixed delays. Experiments on the LibriSpeech benchmark show that ESTRA achieves streaming performance comparable to offline AED models, surpasses local agreement in robustness under small input windows, and reduces 50th-percentile latency by avoiding fixed window-size delays, while leaving room for improvement at the 90th percentile. Furthermore, it provides more reliable control of token emission than probability- or entropy-based baselines, demonstrating the effectiveness of uncertainty as an indicator. ESTRA offers a promising approach to streaming ASR, with results supporting the effectiveness of uncertainty-driven token emission.
基于注意力的编码器-解码器(AED)模型在离线自动语音识别(ASR)中具有较高的准确性,但由于缺乏调节令牌发射的机制,其在流媒体中的应用仍然具有挑战性。现有的方法包括单调注意、与提供标记级边界的外部模型强制对齐以及基于编码器的发射控制方法。然而,这些方法要么需要结构修改,使训练管道复杂化,要么显示有限的准确性。此外,局部协议已经被提出作为一种无需再训练的流式传输方法,但它会产生与输入窗口大小和过早承诺相对应的固定延迟。为了解决这些限制,我们提出了证据流转换器(ESTRA),这是一个利用证据深度学习(EDL)来估计不确定性的框架。ESTRA用Dirichlet分布对令牌概率进行建模,并引入分层和直接的Kullback-Leibler散度损失,以确保随着观察到的语音越来越多,不确定性逐渐降低。在推理过程中,通过将不确定性与阈值进行比较来控制令牌发射,从而抑制没有固定延迟的过早输出。在LibriSpeech基准上的实验表明,ESTRA实现了与离线AED模型相当的流性能,在小输入窗口下的鲁棒性超过了本地协议,并通过避免固定窗口大小的延迟减少了50百分位数的延迟,同时在第90百分位数留下了改进的空间。此外,它提供了比基于概率或熵的基线更可靠的代币发行控制,证明了不确定性作为指标的有效性。ESTRA提供了一种有前途的流式ASR方法,其结果支持不确定性驱动的令牌发射的有效性。
{"title":"Uncertainty-Based Streaming ASR With Evidential Deep Learning","authors":"Hiroaki Sato;Asahi Sakuma;Ryuga Sugano;Tadashi Kumano;Yoshihiko Kawai;Shinji Watanabe;Tetsuji Ogawa","doi":"10.1109/OJSP.2026.3657308","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3657308","url":null,"abstract":"Attention-based encoder-decoder (AED) models achieve high accuracy in offline automatic speech recognition (ASR), but their application to streaming remains challenging due to the lack of mechanisms for regulating token emission. Existing approaches include monotonic attention, forced alignment with external models providing token-level boundaries, and encoder-based emission control methods. However, these methods either require structural modifications, complicate the training pipeline, or show limited accuracy. In addition, local agreement has been proposed as a method enabling streaming without retraining, but it incurs fixed delays corresponding to the input window size and premature commitments. To address these limitations, we propose Evidential Streaming TRAnsformer (ESTRA), a framework that leverages evidential deep learning (EDL) to estimate uncertainty. ESTRA models token probabilities with a Dirichlet distribution and introduces hierarchical and direct Kullback–Leibler divergence losses to ensure uncertainty decreases progressively as more speech is observed. During inference, token emission is controlled by comparing uncertainty against a threshold, suppressing premature outputs without fixed delays. Experiments on the LibriSpeech benchmark show that ESTRA achieves streaming performance comparable to offline AED models, surpasses local agreement in robustness under small input windows, and reduces 50th-percentile latency by avoiding fixed window-size delays, while leaving room for improvement at the 90th percentile. Furthermore, it provides more reliable control of token emission than probability- or entropy-based baselines, demonstrating the effectiveness of uncertainty as an indicator. ESTRA offers a promising approach to streaming ASR, with results supporting the effectiveness of uncertainty-driven token emission.","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"373-381"},"PeriodicalIF":2.7,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11362987","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"147440652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Spherical Acoustic Spatial Entropy: Predicting Acoustic Scene Complexity in Virtual Environments 球形声学空间熵:预测虚拟环境中的声学场景复杂性
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-23 DOI: 10.1109/OJSP.2026.3657297
Luca Resti;Amelia Gully;Michael McLoughlin;Gavin Kearney;Alena Denisova
The objective quantification of Acoustic Scene Complexity (ASC) remains a significant challenge. While existing entropy-based metrics capture spectro-temporal variability, a metric accounting for the spatial distribution of sources has been lacking. We introduce Spherical Acoustic Spatial Entropy (SASE), a novel information-theoretic metric designed for Virtual Acoustic Environments (VAEs). SASE leverages ground-truth spatial data, utilizing an equal-area spherical partition around the listener and weighting source contributions by their perceptual loudness (ITU-R BS.1770-4). We validated SASE through a psychoacoustic experiment (N=21) using a $2times 2 times 2$ factorial design that manipulated masker count, spatial distribution, and motion. SASE was evaluated alongside energy and spectral entropy metrics against subjective ratings of complexity, effort, and spatial spread. Results show that SASE mean was the most robust predictor of perceived complexity in condition-level ratings ($R^{2}=0.714$, $p=0.008$), outperforming spectral and energy entropy. A random-effects pooling of participant regression coefficients confirmed this relationship at the population level ($R^{2}_{text{ {pseudo}}}=0.740$, $p< .001$). Furthermore, a model combining SASE mean with spectral entropy standard deviation explained 84.4% of the variance in perceived complexity, indicating spatial and spectro-temporal metrics capture complementary scene dynamics. SASE provides an objective measure of spatial complexity, enhancing existing frameworks for predicting ASC in virtual environments.
声场景复杂性(ASC)的客观量化仍然是一个重大挑战。虽然现有的基于熵的度量捕获光谱-时间变异性,但缺乏考虑源空间分布的度量。本文介绍了一种新的用于虚拟声环境(VAEs)的信息理论度量——球面声空间熵(SASE)。SASE利用地面真值空间数据,利用听众周围的等面积球形分区,并根据其感知响度加权源的贡献(ITU-R BS.1770-4)。我们通过心理声学实验(N=21)验证了SASE,使用2 × 2 × 2$因子设计,控制掩模计数、空间分布和运动。SASE与能量和谱熵指标一起评估,以对抗复杂性、工作量和空间扩展的主观评级。结果表明,SASE均值是状态水平评分感知复杂性的最稳健预测因子($R^{2}=0.714$, $p=0.008$),优于光谱和能量熵。参与者回归系数的随机效应池在总体水平上证实了这种关系($R^{2}_{text{{pseudo}}}=0.740$, $p< .001$)。此外,将SASE均值与光谱熵标准差相结合的模型解释了感知复杂性方差的84.4%,表明空间和光谱-时间度量捕获了互补的场景动态。SASE提供了空间复杂性的客观度量,增强了虚拟环境中预测ASC的现有框架。
{"title":"Spherical Acoustic Spatial Entropy: Predicting Acoustic Scene Complexity in Virtual Environments","authors":"Luca Resti;Amelia Gully;Michael McLoughlin;Gavin Kearney;Alena Denisova","doi":"10.1109/OJSP.2026.3657297","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3657297","url":null,"abstract":"The objective quantification of Acoustic Scene Complexity (ASC) remains a significant challenge. While existing entropy-based metrics capture spectro-temporal variability, a metric accounting for the spatial distribution of sources has been lacking. We introduce Spherical Acoustic Spatial Entropy (SASE), a novel information-theoretic metric designed for Virtual Acoustic Environments (VAEs). SASE leverages ground-truth spatial data, utilizing an equal-area spherical partition around the listener and weighting source contributions by their perceptual loudness (ITU-R BS.1770-4). We validated SASE through a psychoacoustic experiment (N=21) using a <inline-formula><tex-math>$2times 2 times 2$</tex-math></inline-formula> factorial design that manipulated masker count, spatial distribution, and motion. SASE was evaluated alongside energy and spectral entropy metrics against subjective ratings of complexity, effort, and spatial spread. Results show that SASE mean was the most robust predictor of perceived complexity in condition-level ratings (<inline-formula><tex-math>$R^{2}=0.714$</tex-math></inline-formula>, <inline-formula><tex-math>$p=0.008$</tex-math></inline-formula>), outperforming spectral and energy entropy. A random-effects pooling of participant regression coefficients confirmed this relationship at the population level (<inline-formula><tex-math>$R^{2}_{text{ {pseudo}}}=0.740$</tex-math></inline-formula>, <inline-formula><tex-math>$p&lt; .001$</tex-math></inline-formula>). Furthermore, a model combining SASE mean with spectral entropy standard deviation explained 84.4% of the variance in perceived complexity, indicating spatial and spectro-temporal metrics capture complementary scene dynamics. SASE provides an objective measure of spatial complexity, enhancing existing frameworks for predicting ASC in virtual environments.","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"144-153"},"PeriodicalIF":2.7,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11362972","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175830","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sci-Phi: A Large Language Model Spatial Audio Descriptor Sci-Phi:一个大型语言模型空间音频描述符
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-23 DOI: 10.1109/OJSP.2026.3657300
Xilin Jiang;Hannes Gamper;Sebastian Braun
Acoustic scene perception involves describing the type of sounds, their timing, their direction and distance, as well as their loudness and reverberation. While audio language models excel in sound recognition, single-channel input fundamentally limits spatial understanding. This work presents Sci-Phi, a spatial audio large language model with dual spatial and spectral encoders that estimates a complete parameter set for all sound sources and the surrounding environment. Learning from over 4,000 hours of synthetic first-order Ambisonics recordings including metadata, Sci-Phi enumerates and describes up to four directional sound sources in one pass, alongside non-directional background sounds and room characteristics. We evaluate the model with a permutation-invariant protocol and 15 metrics covering content, location, timing, loudness, and reverberation, and analyze its robustness across source counts, signal-to-noise ratios, reverberation levels, and challenging mixtures of acoustically, spatially, or temporally similar sources. Notably, Sci-Phi generalizes to real room impulse responses with only minor performance degradation. Overall, this work establishes the first audio LLM capable of full spatial-scene description, with strong potential for real-world deployment. Demo: https://sci-phi-audio.github.io/demo
声音场景感知包括描述声音的类型、时间、方向和距离,以及它们的响度和混响。虽然音频语言模型在声音识别方面表现出色,但单通道输入从根本上限制了空间理解。这项工作提出了Sci-Phi,一个空间音频大语言模型,具有双空间和频谱编码器,可以估计所有声源和周围环境的完整参数集。从4000多个小时的合成一阶立体声录音中学习,包括元数据,Sci-Phi在一个通道中枚举和描述多达四个定向声源,以及非定向背景声音和房间特征。我们使用排列不变协议和涵盖内容、位置、时间、响度和混响的15个指标来评估该模型,并分析其在声源计数、信噪比、混响水平以及声学、空间或时间相似声源的挑战性混合中的稳健性。值得注意的是,Sci-Phi推广到真实的房间脉冲响应,只有轻微的性能下降。总的来说,这项工作建立了第一个能够进行完整空间场景描述的音频LLM,具有在现实世界中部署的强大潜力。演示:https://sci-phi-audio.github.io/demo
{"title":"Sci-Phi: A Large Language Model Spatial Audio Descriptor","authors":"Xilin Jiang;Hannes Gamper;Sebastian Braun","doi":"10.1109/OJSP.2026.3657300","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3657300","url":null,"abstract":"Acoustic scene perception involves describing the type of sounds, their timing, their direction and distance, as well as their loudness and reverberation. While audio language models excel in sound recognition, single-channel input fundamentally limits spatial understanding. This work presents <italic>Sci-Phi</i>, a spatial audio large language model with dual spatial and spectral encoders that estimates a complete parameter set for all sound sources and the surrounding environment. Learning from over 4,000 hours of synthetic first-order Ambisonics recordings including metadata, <italic>Sci-Phi</i> enumerates and describes up to four directional sound sources in one pass, alongside non-directional background sounds and room characteristics. We evaluate the model with a permutation-invariant protocol and 15 metrics covering content, location, timing, loudness, and reverberation, and analyze its robustness across source counts, signal-to-noise ratios, reverberation levels, and challenging mixtures of acoustically, spatially, or temporally similar sources. Notably, <italic>Sci-Phi</i> generalizes to real room impulse responses with only minor performance degradation. Overall, this work establishes the first audio LLM capable of full spatial-scene description, with strong potential for real-world deployment. Demo: <uri>https://sci-phi-audio.github.io/demo</uri>","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"276-284"},"PeriodicalIF":2.7,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11362973","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146223560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributionally Robust Ultra-Reliable Resource Allocation via Double Tail Waterfilling Under Fading Risk 衰落风险下双尾注水的分布鲁棒超可靠资源分配
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-23 DOI: 10.1109/OJSP.2026.3657284
Gokberk Yaylali;Dionysis Kalogerias
Optimal resource allocation in wireless systems remains a fundamental challenge due to the inherent adversities caused by channel fading.Modern wireless applications require efficient allocation schemes that maximize total network utility ensuring robust and reliable system performance. Although optimal on average, ergodic-optimal policies, commonly realized via stochastic waterfilling schemes, are susceptible to statistical dispersion of commonly heavy-tailed or highly volatile fading channels, particularly in terms of both instantaneous power policy fluctuations and frequent service outages (due to deep fade events), violating established power-level and quality-of-service specifications, essentially sabotaging fulfillment of provider-specific power/energy targets on the one hand, and user-perceived system reliability on the other. At the other extreme, short-term-optimal policies, commonly relying on deterministic waterfilling, or maximally averse minimax-optimal policies, strictly satisfy specifications but are computationally demanding, impractical, while also being suboptimal in any long-term regime. To address these challenges, we introduce a distributionally robust formulation of the constrained stochastic resource allocation problem in the classical point-to-point interference-free multi-terminal network by leveraging Conditional Value-at-Risk (CVaR) as a coherent measure of fading and/or fluctuation risk relevant to both transmission power and achievable rate distributions. We derive a closed-form parameterized expression for the CVaR-optimal resource policy which is of remarkably simple and interpretable form, along with subgradient-based update schemes for the corresponding CVaR quantile levels to both transmission power and achievable rates. Building on this, we develop a primal-dual double tail waterfilling scheme which iteratively computes globally optimal policies achieving ultra-reliable long-term rate performance, but with near-short-term characteristics. Extensive numerical experiments corroborate the effectiveness of the proposed approach.
由于信道衰落带来的固有弊端,无线系统的资源优化分配仍然是一个根本性的挑战。现代无线应用需要有效的分配方案,以最大限度地提高总网络效用,确保系统性能的鲁棒性和可靠性。虽然平均而言是最优的,但遍历最优策略(通常通过随机注水方案实现)容易受到通常重尾或高度波动的衰落信道的统计分散的影响,特别是在电力政策的瞬时波动和频繁的服务中断(由于深度衰落事件)方面,违反了既定的功率水平和服务质量规范。本质上一方面破坏了供应商特定的功率/能量目标的实现,另一方面破坏了用户感知的系统可靠性。在另一个极端,短期最优策略,通常依赖于确定性充水,或最大厌恶最小最优策略,严格满足规范,但计算要求高,不切实际,同时在任何长期制度下也是次优的。为了解决这些挑战,我们引入了经典点对点无干扰多终端网络中约束随机资源分配问题的分布鲁棒性公式,利用条件风险值(CVaR)作为与传输功率和可实现速率分布相关的衰落和/或波动风险的一致度量。我们推导了CVaR最优资源策略的封闭参数化表达式,该表达式具有非常简单和可解释的形式,以及相应的CVaR分位数水平对传输功率和可实现速率的基于子梯度的更新方案。在此基础上,我们开发了一种原始-双重双尾注水方案,该方案迭代计算全局最优策略,实现超可靠的长期费率性能,但具有近短期特征。大量的数值实验证实了该方法的有效性。
{"title":"Distributionally Robust Ultra-Reliable Resource Allocation via Double Tail Waterfilling Under Fading Risk","authors":"Gokberk Yaylali;Dionysis Kalogerias","doi":"10.1109/OJSP.2026.3657284","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3657284","url":null,"abstract":"Optimal resource allocation in wireless systems remains a fundamental challenge due to the inherent adversities caused by channel fading.Modern wireless applications require efficient allocation schemes that maximize total network utility ensuring robust and reliable system performance. Although optimal on average, ergodic-optimal policies, commonly realized via stochastic waterfilling schemes, are susceptible to statistical dispersion of commonly heavy-tailed or highly volatile fading channels, particularly in terms of <italic>both</i> instantaneous power policy fluctuations <italic>and</i> frequent service outages (due to deep fade events), violating established power-level and quality-of-service specifications, essentially sabotaging fulfillment of provider-specific power/energy targets on the one hand, and user-perceived system reliability on the other. At the other extreme, short-term-optimal policies, commonly relying on deterministic waterfilling, or maximally averse minimax-optimal policies, strictly satisfy specifications but are computationally demanding, impractical, while also being suboptimal in any long-term regime. To address these challenges, we introduce a distributionally robust formulation of the constrained stochastic resource allocation problem in the classical point-to-point interference-free multi-terminal network by leveraging Conditional Value-at-Risk (CVaR) as a coherent measure of fading and/or fluctuation risk relevant to both transmission power and achievable rate distributions. We derive a closed-form parameterized expression for the CVaR-optimal resource policy which is of remarkably simple and interpretable form, along with subgradient-based update schemes for the corresponding CVaR quantile levels to both transmission power and achievable rates. Building on this, we develop a primal-dual <italic>double tail waterfilling</i> scheme which iteratively computes <italic>globally optimal policies achieving ultra-reliable long-term rate performance, but with near-short-term characteristics</i>. Extensive numerical experiments corroborate the effectiveness of the proposed approach.","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"154-164"},"PeriodicalIF":2.7,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11362910","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175827","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Leveraging Beam Search Information for Confidence Estimation in E2E ASR 利用波束搜索信息估计E2E ASR的置信度
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-23 DOI: 10.1109/OJSP.2026.3657286
Yichen Jia;Hugo Van hamme
To estimate confidence for end-to-end Automatic Speech Recognition (ASR) systems, recent research has proposed Confidence Estimation Modules that incorporate features from the backbone ASR model. Most existing approaches, however, are architecture-dependent. In this paper, we propose the Score-Rank Confidence Estimation Module (SR-CEM), a lightweight module that leverages beam search information to generate token- and word-level confidence scores. Specifically, SR-CEM constructs features by combining the scores and ranks of tokens within a hypothesis. Experiments show that SR-CEM achieves effective calibration on both in-domain and out-of-domain English data. On the in-domain test set, it attains a Maximum Calibration Error of 4.50% and an Expected Calibration Error of 0.30% at the token level, significantly outperforming softmax confidence (20.04% and 1.75%, respectively). At the word level, SR-CEM achieves 8.17% and 0.35%, compared to 17.91% and 1.67% from softmax confidence. Furthermore, we demonstrate its robustness across hybrid and transducer ASR architectures with different decoding strategies, as well as on Dutch, noisy and conversational speech conditions. Our main finding is that SR-CEM is particularly effective in reducing Maximum Calibration Error, which is critical for reliable downstream use of ASR outputs, while maintaining architecture independence and generality across diverse evaluation conditions.
为了估计端到端自动语音识别(ASR)系统的置信度,最近的研究提出了包含骨干语音识别模型特征的置信度估计模块。然而,大多数现有的方法是依赖于体系结构的。在本文中,我们提出了分数-排名置信度估计模块(SR-CEM),这是一个轻量级模块,它利用波束搜索信息来生成令牌和词级置信度评分。具体来说,SR-CEM通过结合假设中的标记的分数和等级来构建特征。实验表明,SR-CEM对域内和域外的英语数据都能实现有效的标定。在域内测试集上,它在令牌水平上获得了4.50%的最大校准误差和0.30%的预期校准误差,显著优于softmax置信度(分别为20.04%和1.75%)。在单词水平上,SR-CEM达到8.17%和0.35%,而softmax置信度为17.91%和1.67%。此外,我们展示了其在具有不同解码策略的混合和换能器ASR架构以及荷兰语,嘈杂和会话语音条件下的鲁棒性。我们的主要发现是SR-CEM在降低最大校准误差方面特别有效,这对于ASR输出的可靠下游使用至关重要,同时在不同的评估条件下保持架构的独立性和通用性。
{"title":"Leveraging Beam Search Information for Confidence Estimation in E2E ASR","authors":"Yichen Jia;Hugo Van hamme","doi":"10.1109/OJSP.2026.3657286","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3657286","url":null,"abstract":"To estimate confidence for end-to-end Automatic Speech Recognition (ASR) systems, recent research has proposed Confidence Estimation Modules that incorporate features from the backbone ASR model. Most existing approaches, however, are architecture-dependent. In this paper, we propose the Score-Rank Confidence Estimation Module (SR-CEM), a lightweight module that leverages beam search information to generate token- and word-level confidence scores. Specifically, SR-CEM constructs features by combining the scores and ranks of tokens within a hypothesis. Experiments show that SR-CEM achieves effective calibration on both in-domain and out-of-domain English data. On the in-domain test set, it attains a Maximum Calibration Error of 4.50% and an Expected Calibration Error of 0.30% at the token level, significantly outperforming softmax confidence (20.04% and 1.75%, respectively). At the word level, SR-CEM achieves 8.17% and 0.35%, compared to 17.91% and 1.67% from softmax confidence. Furthermore, we demonstrate its robustness across hybrid and transducer ASR architectures with different decoding strategies, as well as on Dutch, noisy and conversational speech conditions. Our main finding is that SR-CEM is particularly effective in reducing Maximum Calibration Error, which is critical for reliable downstream use of ASR outputs, while maintaining architecture independence and generality across diverse evaluation conditions.","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"125-133"},"PeriodicalIF":2.7,"publicationDate":"2026-01-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11362960","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175829","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Fully Complex-Valued Underwater Acoustic Signal Enhancement Model for Passive Sonar Systems 被动声呐系统的全复值水声信号增强模型
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-20 DOI: 10.1109/OJSP.2026.3656063
Zhengzhe Zhang;Jie Zhang;Haoyin Yan;Hengshuang Liu;Junhua Liu
Passive sonar systems offer stealth and low-energy consumption advantages while facing highly complex signal conditions. Underwater acoustic signal (UWAS) enhancement for passive sonar systems aims to improve the quality of vessel-radiated signals captured by hydrophones, thereby facilitating subsequent tasks like target recognition. However, conventional methods might struggle due to the intricate marine environment and weak target signals. In this work, we propose a fully Complex-valued U-Net based Multidimensional Attention Network (CUMA-Net), with all modules operating in the complex domain to jointly exploit magnitude and phase information. CUMA-Net employs a complex-valued encoder-decoder, which captures multiscale features for spectral mapping. To boost representation power and emphasize line spectrum components, we incorporate a complex-valued multidimensional attention module. This module includes a complex-valued time-frequency conformer to model dependencies along temporal and frequency axes. Complementarily, a complex convolutional block attention module extracts features across spatial and channel dimensions. To guide training under low SNR conditions, we propose a normalized mean squared error loss tailored for spectrogram reconstruction. Results on a public dataset verify that CUMA-Net achieves superior UWAS enhancement performance, while the improved signal quality further benefits vessel classification. Furthermore, we explore the impact of input frequency resolution on both enhancement and classification performance.
被动声呐系统在面对高度复杂的信号条件时具有隐身和低能耗优势。被动声呐系统的水声信号增强旨在提高水听器捕获的船舶辐射信号的质量,从而促进后续任务,如目标识别。然而,由于复杂的海洋环境和微弱的目标信号,传统的方法可能会遇到困难。在这项工作中,我们提出了一个基于全复值U-Net的多维注意力网络(CUMA-Net),所有模块都在复杂域中工作,以共同利用幅度和相位信息。CUMA-Net采用复值编码器-解码器,捕获多尺度特征进行光谱映射。为了提高表征能力和强调线谱分量,我们加入了一个复杂值多维关注模块。该模块包括一个复值时频共形器,用于沿时间轴和频率轴建模依赖关系。此外,一个复杂的卷积块注意力模块提取跨空间和通道维度的特征。为了指导低信噪比条件下的训练,我们提出了为谱图重建量身定制的归一化均方误差损失。公共数据集上的结果验证了CUMA-Net实现了卓越的UWAS增强性能,而改进的信号质量进一步有利于船舶分类。此外,我们探讨了输入频率分辨率对增强和分类性能的影响。
{"title":"A Fully Complex-Valued Underwater Acoustic Signal Enhancement Model for Passive Sonar Systems","authors":"Zhengzhe Zhang;Jie Zhang;Haoyin Yan;Hengshuang Liu;Junhua Liu","doi":"10.1109/OJSP.2026.3656063","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3656063","url":null,"abstract":"Passive sonar systems offer stealth and low-energy consumption advantages while facing highly complex signal conditions. Underwater acoustic signal (UWAS) enhancement for passive sonar systems aims to improve the quality of vessel-radiated signals captured by hydrophones, thereby facilitating subsequent tasks like target recognition. However, conventional methods might struggle due to the intricate marine environment and weak target signals. In this work, we propose a fully Complex-valued U-Net based Multidimensional Attention Network (CUMA-Net), with all modules operating in the complex domain to jointly exploit magnitude and phase information. CUMA-Net employs a complex-valued encoder-decoder, which captures multiscale features for spectral mapping. To boost representation power and emphasize line spectrum components, we incorporate a complex-valued multidimensional attention module. This module includes a complex-valued time-frequency conformer to model dependencies along temporal and frequency axes. Complementarily, a complex convolutional block attention module extracts features across spatial and channel dimensions. To guide training under low SNR conditions, we propose a normalized mean squared error loss tailored for spectrogram reconstruction. Results on a public dataset verify that CUMA-Net achieves superior UWAS enhancement performance, while the improved signal quality further benefits vessel classification. Furthermore, we explore the impact of input frequency resolution on both enhancement and classification performance.","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"101-115"},"PeriodicalIF":2.7,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11359482","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning-Based Event Data Coding: A Joint Spatiotemporal and Polarity Solution 基于深度学习的事件数据编码:一种时空和极性联合解决方案
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-20 DOI: 10.1109/OJSP.2026.3656104
Abdelrahman Seleem;André F. R. Guarda;Nuno M. M. Rodrigues;Fernando Pereira
Neuromorphic vision sensors, commonly referred to as event cameras, generate a massive number of pixel-level events, composed by spatiotemporal and polarity information, thus demanding highly efficient coding solutions. Existing solutions focus on lossless coding of event data, assuming that no distortion is acceptable for the target use cases, mostly including computer vision tasks such as classification and recognition. One promising coding approach exploits the similarity between event data and point clouds, both being sets of 3D points, thus allowing to use current point cloud coding solutions to code event data, typically adopting a two-point clouds representation, one for each event polarity. This paper proposes a novel lossy Deep Learning-based Joint Event data Coding (DL-JEC) solution, which adopts for the first time a single-point cloud representation, where the event polarity plays the role of a point cloud attribute, thus enabling to exploit the correlation between the geometry/spatiotemporal and polarity event information. Moreover, this paper also proposes novel adaptive voxel binarization strategies which may be used in DL-JEC, optimized for either quality-oriented or computer vision task-oriented purposes which allow to maximize the performance for the task at hand. DL-JEC can achieve significant compression performance gains when compared with relevant conventional and DL-based state-of-the-art event data coding solutions, notably the MPEG G-PCC and JPEG Pleno PCC standards. Furthermore, it is shown that it is possible to use lossy event data coding, with significantly reduced rate regarding lossless coding, without compromising the target computer vision task performance, notably event classification, thus changing the current event data coding paradigm.
神经形态视觉传感器通常被称为事件相机,它产生大量由时空和极性信息组成的像素级事件,因此需要高效的编码解决方案。现有的解决方案侧重于事件数据的无损编码,假设目标用例不能接受失真,主要包括分类和识别等计算机视觉任务。一种有前途的编码方法利用事件数据和点云之间的相似性,两者都是3D点的集合,因此允许使用当前的点云编码解决方案来编码事件数据,通常采用两个点云表示,每个点云表示一个事件极性。本文提出了一种新的基于有损深度学习的联合事件数据编码(DL-JEC)解决方案,该方案首次采用单点云表示,其中事件极性扮演点云属性的角色,从而能够利用几何/时空和极性事件信息之间的相关性。此外,本文还提出了新的自适应体素二值化策略,可用于DL-JEC,针对面向质量或面向计算机视觉任务的目的进行优化,从而最大限度地提高手头任务的性能。与相关的传统和基于dl的最先进的事件数据编码解决方案(特别是MPEG G-PCC和JPEG Pleno PCC标准)相比,DL-JEC可以实现显著的压缩性能提升。此外,研究表明,在不影响目标计算机视觉任务性能(特别是事件分类)的情况下,使用有损事件数据编码可以显著降低无损编码的速率,从而改变当前的事件数据编码范式。
{"title":"Deep Learning-Based Event Data Coding: A Joint Spatiotemporal and Polarity Solution","authors":"Abdelrahman Seleem;André F. R. Guarda;Nuno M. M. Rodrigues;Fernando Pereira","doi":"10.1109/OJSP.2026.3656104","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3656104","url":null,"abstract":"Neuromorphic vision sensors, commonly referred to as event cameras, generate a massive number of pixel-level events, composed by spatiotemporal and polarity information, thus demanding highly efficient coding solutions. Existing solutions focus on lossless coding of event data, assuming that no distortion is acceptable for the target use cases, mostly including computer vision tasks such as classification and recognition. One promising coding approach exploits the similarity between event data and point clouds, both being sets of 3D points, thus allowing to use current point cloud coding solutions to code event data, typically adopting a two-point clouds representation, one for each event polarity. This paper proposes a novel lossy Deep Learning-based Joint Event data Coding (DL-JEC) solution, which adopts for the first time a single-point cloud representation, where the event polarity plays the role of a point cloud attribute, thus enabling to exploit the correlation between the geometry/spatiotemporal and polarity event information. Moreover, this paper also proposes novel adaptive voxel binarization strategies which may be used in DL-JEC, optimized for either quality-oriented or computer vision task-oriented purposes which allow to maximize the performance for the task at hand. DL-JEC can achieve significant compression performance gains when compared with relevant conventional and DL-based state-of-the-art event data coding solutions, notably the MPEG G-PCC and JPEG Pleno PCC standards. Furthermore, it is shown that it is possible to use lossy event data coding, with significantly reduced rate regarding lossless coding, without compromising the target computer vision task performance, notably event classification, thus changing the current event data coding paradigm.","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"222-237"},"PeriodicalIF":2.7,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11359485","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175986","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Test-Time Adaptation for Speech Enhancement via Domain Invariant Embedding Transformation 基于域不变嵌入变换的语音增强测试时间自适应
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-20 DOI: 10.1109/OJSP.2026.3656059
Tobias Raichle;Niels Edinger;Bin Yang
Deep learning-based speech enhancement models achieve remarkable performance when test distributions match training conditions, but often degrade when deployed in unpredictable real-world environments with domain shifts. To address this challenge, we present laden, the first test-time adaptation method specifically designed for speech enhancement. Our approach leverages powerful pre-trained speech representations to perform latent denoising, approximating clean speech representations through a linear transformation of noisy embeddings. We show that this transformation generalizes well across domains, enabling effective pseudo-labeling for target domains without labeled target data. The resulting pseudo-labels enable effective test-time adaptation of speech enhancement models across diverse acoustic environments. We propose a comprehensive benchmark spanning multiple datasets with various domain shifts, including changes in noise types, speaker characteristics, and languages. Our extensive experiments demonstrate that LaDen consistently outperforms baseline methods across perceptual metrics, particularly for speaker and language domain shifts.
当测试分布与训练条件相匹配时,基于深度学习的语音增强模型取得了显著的性能,但当部署在不可预测的现实世界环境中并伴有域转移时,其性能往往会下降。为了解决这一挑战,我们提出了第一种专门为语音增强设计的测试时间自适应方法。我们的方法利用强大的预训练语音表示来执行潜在去噪,通过噪声嵌入的线性变换近似干净的语音表示。我们证明了这种转换可以很好地跨域推广,从而可以在没有标记目标数据的情况下对目标域进行有效的伪标记。由此产生的伪标签能够有效地适应不同声学环境下的语音增强模型的测试时间。我们提出了一个涵盖多个数据集的综合基准,这些数据集具有不同的域位移,包括噪声类型、说话者特征和语言的变化。我们的大量实验表明,LaDen在感知指标上的表现始终优于基线方法,特别是在说话者和语言域转移方面。
{"title":"Test-Time Adaptation for Speech Enhancement via Domain Invariant Embedding Transformation","authors":"Tobias Raichle;Niels Edinger;Bin Yang","doi":"10.1109/OJSP.2026.3656059","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3656059","url":null,"abstract":"Deep learning-based speech enhancement models achieve remarkable performance when test distributions match training conditions, but often degrade when deployed in unpredictable real-world environments with domain shifts. To address this challenge, we present laden, the first test-time adaptation method specifically designed for speech enhancement. Our approach leverages powerful pre-trained speech representations to perform latent denoising, approximating clean speech representations through a linear transformation of noisy embeddings. We show that this transformation generalizes well across domains, enabling effective pseudo-labeling for target domains without labeled target data. The resulting pseudo-labels enable effective test-time adaptation of speech enhancement models across diverse acoustic environments. We propose a comprehensive benchmark spanning multiple datasets with various domain shifts, including changes in noise types, speaker characteristics, and languages. Our extensive experiments demonstrate that LaDen consistently outperforms baseline methods across perceptual metrics, particularly for speaker and language domain shifts.","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"134-143"},"PeriodicalIF":2.7,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11359505","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175950","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design of Acoustic Equalization Filters for Headphones Based on Low-Rank Regularization 基于低秩正则化的耳机声学均衡滤波器设计
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-20 DOI: 10.1109/OJSP.2026.3656057
Florian Hilgemann;Peter Jax
The use of equalization filters to achieve acoustic transparency can improve the sound quality of hearables and hearing aids. Finite impulse response (FIR) filters guarantee stability and offer a listening impression close to the open ear, but their implementation may conflict with the resource constraints typical of hearing devices. Infinite impulse response (IIR) filters are commonly used to meet these constraints, but their design often lacks stability and performance guarantees. Therefore, we consider indirect IIR filter design methods that extend FIR filter designs with an IIR approximation step. To mitigate the performance degradation caused by the IIR approximation, we establish a formal connection between optimization variable and IIR approximation error, and propose an approximation-aware design algorithm based on the nuclear norm heuristic. The evaluation considers the design of hear-through filters using real-world measurement data. The proposed approach can reduce the time-domain mean-squared error by up to $text{6},text{dB}$ compared to conventional methods, and shows a high robustness against between-person variance. Thus, the results offer an improvement in hearing device personalization within practical constraints.
利用均衡滤波器实现声透明,可以提高耳机和助听器的音质。有限脉冲响应(FIR)滤波器保证了稳定性,并提供靠近开放耳朵的听音印象,但它们的实现可能与听力设备典型的资源限制相冲突。无限脉冲响应(IIR)滤波器通常用于满足这些约束,但其设计往往缺乏稳定性和性能保证。因此,我们考虑间接IIR滤波器设计方法,通过IIR近似步骤扩展FIR滤波器设计。为了减轻IIR近似引起的性能下降,建立了优化变量与IIR近似误差之间的形式化联系,提出了一种基于核范数启发式的逼近感知设计算法。评估考虑了使用实际测量数据的透听滤波器的设计。与传统方法相比,该方法可将时域均方误差降低高达$text{6},text{dB}$,并且对人间方差具有较高的鲁棒性。因此,该结果在实际限制下提供了助听器个性化的改进。
{"title":"Design of Acoustic Equalization Filters for Headphones Based on Low-Rank Regularization","authors":"Florian Hilgemann;Peter Jax","doi":"10.1109/OJSP.2026.3656057","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3656057","url":null,"abstract":"The use of equalization filters to achieve acoustic transparency can improve the sound quality of hearables and hearing aids. Finite impulse response (FIR) filters guarantee stability and offer a listening impression close to the open ear, but their implementation may conflict with the resource constraints typical of hearing devices. Infinite impulse response (IIR) filters are commonly used to meet these constraints, but their design often lacks stability and performance guarantees. Therefore, we consider indirect IIR filter design methods that extend FIR filter designs with an IIR approximation step. To mitigate the performance degradation caused by the IIR approximation, we establish a formal connection between optimization variable and IIR approximation error, and propose an approximation-aware design algorithm based on the nuclear norm heuristic. The evaluation considers the design of hear-through filters using real-world measurement data. The proposed approach can reduce the time-domain mean-squared error by up to <inline-formula><tex-math>$text{6},text{dB}$</tex-math></inline-formula> compared to conventional methods, and shows a high robustness against between-person variance. Thus, the results offer an improvement in hearing device personalization within practical constraints.","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"173-184"},"PeriodicalIF":2.7,"publicationDate":"2026-01-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11359448","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-Blind Channel Estimation for Single-Carrier With Frequency-Domain Equalization Systems in the Presence of Colored Noise : A Cyclic Zero Insertion Approach 彩色噪声存在下单载波频域均衡系统的半盲信道估计:循环零插入方法
IF 2.7 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-01-19 DOI: 10.1109/OJSP.2026.3655219
Yi-Sheng Chen;Hung-Shuo Chang
In this paper, we propose a semi-blind channel estimation method for single-carrier with frequency-domain equalization (SC-FDE) systems in the presence of colored noise. The method introduces a cyclic zero insertion approach where a single zero is periodically inserted into the transmitted sequence. This way induces a special structure of the autocorrelation matrix of the received signal, enabling effective channel estimation even in environments with colored noise. By extracting the channel product coefficients from the autocorrelation matrix, we construct a Hermitian matrix whose dominant eigenvector corresponds to the channel impulse response. Simulation results demonstrate the effectiveness of the proposed method.
在本文中,我们提出了一种半盲信道估计方法,用于存在彩色噪声的单载波频域均衡系统。该方法引入了循环零插入方法,其中周期性地将单个零插入到传输序列中。这种方法诱导了接收信号的自相关矩阵的特殊结构,即使在有彩色噪声的环境中也能有效地估计信道。通过从自相关矩阵中提取信道积系数,构造了一个厄米矩阵,其主特征向量对应于信道脉冲响应。仿真结果验证了该方法的有效性。
{"title":"Semi-Blind Channel Estimation for Single-Carrier With Frequency-Domain Equalization Systems in the Presence of Colored Noise : A Cyclic Zero Insertion Approach","authors":"Yi-Sheng Chen;Hung-Shuo Chang","doi":"10.1109/OJSP.2026.3655219","DOIUrl":"https://doi.org/10.1109/OJSP.2026.3655219","url":null,"abstract":"In this paper, we propose a semi-blind channel estimation method for single-carrier with frequency-domain equalization (SC-FDE) systems in the presence of colored noise. The method introduces a cyclic zero insertion approach where a single zero is periodically inserted into the transmitted sequence. This way induces a special structure of the autocorrelation matrix of the received signal, enabling effective channel estimation even in environments with colored noise. By extracting the channel product coefficients from the autocorrelation matrix, we construct a Hermitian matrix whose dominant eigenvector corresponds to the channel impulse response. Simulation results demonstrate the effectiveness of the proposed method.","PeriodicalId":73300,"journal":{"name":"IEEE open journal of signal processing","volume":"7 ","pages":"213-221"},"PeriodicalIF":2.7,"publicationDate":"2026-01-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=11358533","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146175947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE open journal of signal processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1