首页 > 最新文献

2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)最新文献

英文 中文
Coded excitation ultrasound: Efficient implementation via frequency domain processing 编码激发超声:通过频域处理有效实现
Almog Lahav, Yuval Ben-Shalom, T. Chernyakova, Yonina C. Eldar
Modern imaging systems use single-carrier short pulses for transducer excitation. The usage of coded signals allowing for pulse compression is known to improve signal-to-noise ratio (SNR), for example in radar and communication. One of the main challenges in applying coded excitation (CE) to medical imaging is frequency dependent attenuation in biological tissues. Previous work overcame this challenge and verified significant improvement in SNR and imaging depth by using an array of transducer elements and applying pulse compression at each element. However, this approach results in a large computational load. A common way of reducing the cost is to apply pulse compression after beamforming, which reduces image quality. In this work we propose a high-quality low cost method for CE imaging by integrating pulse compression into the recently developed frequency domain beamforming framework. This approach yields a 26-fold reduction in computational complexity without compromising image quality. This reduction enables efficient implementation of CE in array imaging paving the way to enhanced SNR, improved imaging depth and higher frame-rate.
现代成像系统使用单载波短脉冲来激励换能器。使用允许脉冲压缩的编码信号可以提高信噪比(SNR),例如在雷达和通信中。将编码激发(CE)应用于医学成像的主要挑战之一是生物组织中的频率依赖性衰减。之前的工作克服了这一挑战,并通过使用一系列换能器元件并对每个元件进行脉冲压缩,验证了信噪比和成像深度的显着改善。然而,这种方法会导致大量的计算负载。降低成本的一种常用方法是在波束形成后进行脉冲压缩,这降低了图像质量。在这项工作中,我们提出了一种高质量的低成本的CE成像方法,将脉冲压缩集成到最近开发的频域波束形成框架中。这种方法可以在不影响图像质量的情况下将计算复杂度降低26倍。这种减少使得阵列成像中有效地实现了CE,为增强信噪比、改善成像深度和更高的帧率铺平了道路。
{"title":"Coded excitation ultrasound: Efficient implementation via frequency domain processing","authors":"Almog Lahav, Yuval Ben-Shalom, T. Chernyakova, Yonina C. Eldar","doi":"10.1109/ICASSP.2016.7471760","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7471760","url":null,"abstract":"Modern imaging systems use single-carrier short pulses for transducer excitation. The usage of coded signals allowing for pulse compression is known to improve signal-to-noise ratio (SNR), for example in radar and communication. One of the main challenges in applying coded excitation (CE) to medical imaging is frequency dependent attenuation in biological tissues. Previous work overcame this challenge and verified significant improvement in SNR and imaging depth by using an array of transducer elements and applying pulse compression at each element. However, this approach results in a large computational load. A common way of reducing the cost is to apply pulse compression after beamforming, which reduces image quality. In this work we propose a high-quality low cost method for CE imaging by integrating pulse compression into the recently developed frequency domain beamforming framework. This approach yields a 26-fold reduction in computational complexity without compromising image quality. This reduction enables efficient implementation of CE in array imaging paving the way to enhanced SNR, improved imaging depth and higher frame-rate.","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117273657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Dual-stage algorithm to identify channels with poor electrode-to-neuron interface in cochlear implant users 人工耳蜗使用者电极-神经元界面差通道的双阶段识别算法
Stefano Cosentino, Lindsay De Vries, Rachel Scheperle, Julie Bierer, R. Carlyon
Users of cochlear implants rely on a number of electrodes to perceive acoustic information. The extent to which their hearing is restored depends on a number of factors including the electrode-to-neuron interface. We describe an approach to detect instances of poor-performing channels based on physiological data known as electrically evoked compound action potentials (ECAPs). The proposed approach - termed Panoramic ECAP ("PECAP") - combines nonlinear optimization stages with different constraints to recover neural activation patterns for all electrodes. Data were obtained from nine cochlear implant subjects and used to run the PECAP tool to identify possible instances of poor-performing channels. Data from one subject revealed a shifted peak ("dead region").
人工耳蜗的使用者依靠许多电极来感知声音信息。他们的听力恢复程度取决于许多因素,包括电极-神经元界面。我们描述了一种基于被称为电诱发复合动作电位(ECAPs)的生理数据来检测表现不佳的通道实例的方法。所提出的方法被称为全景ECAP(“PECAP”),它结合了具有不同约束条件的非线性优化阶段,以恢复所有电极的神经激活模式。数据来自9名人工耳蜗受试者,并用于运行PECAP工具来识别可能表现不佳的通道。一名受试者的数据显示了一个移位的峰值(“死区”)。
{"title":"Dual-stage algorithm to identify channels with poor electrode-to-neuron interface in cochlear implant users","authors":"Stefano Cosentino, Lindsay De Vries, Rachel Scheperle, Julie Bierer, R. Carlyon","doi":"10.1109/ICASSP.2016.7471792","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7471792","url":null,"abstract":"Users of cochlear implants rely on a number of electrodes to perceive acoustic information. The extent to which their hearing is restored depends on a number of factors including the electrode-to-neuron interface. We describe an approach to detect instances of poor-performing channels based on physiological data known as electrically evoked compound action potentials (ECAPs). The proposed approach - termed Panoramic ECAP (\"PECAP\") - combines nonlinear optimization stages with different constraints to recover neural activation patterns for all electrodes. Data were obtained from nine cochlear implant subjects and used to run the PECAP tool to identify possible instances of poor-performing channels. Data from one subject revealed a shifted peak (\"dead region\").","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121075568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Multichannel audio declipping 多声道音频衰减
A. Ozerov, Ç. Bilen, P. Pérez
Audio declipping consists in recovering so-called clipped audio samples that are set to a maximum / minimum threshold. Many different approaches were proposed to solve this problem in case of singlechannel (mono) recordings. However, while most of audio recordings are multichannel nowadays, there is no method designed specifically for multichannel audio declipping, where the inter-channel correlations may be efficiently exploited for a better declipping result. In this work we propose for the first time such a multichannel audio declipping method. Our method is based on representing a multichannel audio recording as a convolutive mixture of several audio sources, and on modeling the source power spectrograms and mixing filters by nonnegative tensor factorization model and full-rank covariance matrices, respectively. A generalized expectation-maximization algorithm is proposed to estimate model parameters. It is shown experimentally that the proposed multichannel audio de-clipping algorithm outperforms in average and in most cases a state-of-the-art single-channel declipping algorithm applied to each channel independently.
音频衰减包括恢复所谓的剪辑音频样本,设置为最大/最小阈值。许多不同的方法被提出来解决这个问题,在单通道(单声道)录音的情况下。然而,虽然现在大多数音频记录都是多通道的,但没有专门为多通道音频衰减设计的方法,可以有效地利用通道间的相关性来获得更好的衰减结果。在这项工作中,我们首次提出了这样一种多通道音频衰减方法。我们的方法基于将多声道音频记录表示为多个音频源的卷积混合,并分别通过非负张量分解模型和全秩协方差矩阵对源功率谱和混合滤波器进行建模。提出了一种广义期望最大化算法来估计模型参数。实验表明,所提出的多通道音频去剪辑算法在平均和大多数情况下优于独立应用于每个通道的最先进的单通道去剪辑算法。
{"title":"Multichannel audio declipping","authors":"A. Ozerov, Ç. Bilen, P. Pérez","doi":"10.1109/ICASSP.2016.7471757","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7471757","url":null,"abstract":"Audio declipping consists in recovering so-called clipped audio samples that are set to a maximum / minimum threshold. Many different approaches were proposed to solve this problem in case of singlechannel (mono) recordings. However, while most of audio recordings are multichannel nowadays, there is no method designed specifically for multichannel audio declipping, where the inter-channel correlations may be efficiently exploited for a better declipping result. In this work we propose for the first time such a multichannel audio declipping method. Our method is based on representing a multichannel audio recording as a convolutive mixture of several audio sources, and on modeling the source power spectrograms and mixing filters by nonnegative tensor factorization model and full-rank covariance matrices, respectively. A generalized expectation-maximization algorithm is proposed to estimate model parameters. It is shown experimentally that the proposed multichannel audio de-clipping algorithm outperforms in average and in most cases a state-of-the-art single-channel declipping algorithm applied to each channel independently.","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127249368","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Complexity reduction of SUMIS MIMO soft detection based on box optimization for large systems 基于盒优化的大型系统SUMIS MIMO软检测复杂度降低
M. Simarro, V. García, F. Martínez-Zaldívar, Alberto González, A. Vidal
A new algorithm called SUMIS-BO is proposed for soft-output MIMO detection. This method is a meaningful improvement of the "Subspace Marginalization with Interference Suppression" (SUMIS) algorithm. It exhibits good performance with reduced complexity and has been evaluated and compared in terms of performance and efficiency with the SUMIS algorithm using different system parameters. Results show that the performance of the SUMIS-BO is similar to the SUMIS algorithm, however its efficiency is improved. The new algorithm is far more efficient than SUMIS, especially with large systems.
提出了一种用于软输出MIMO检测的SUMIS-BO算法。该方法是对“子空间边缘干扰抑制”(SUMIS)算法的改进。该算法具有较好的性能和较低的复杂度,并与使用不同系统参数的SUMIS算法在性能和效率方面进行了评估和比较。结果表明,SUMIS- bo算法的性能与SUMIS算法相当,但效率有所提高。新算法比SUMIS要高效得多,尤其是在大型系统中。
{"title":"Complexity reduction of SUMIS MIMO soft detection based on box optimization for large systems","authors":"M. Simarro, V. García, F. Martínez-Zaldívar, Alberto González, A. Vidal","doi":"10.1109/ICASSP.2016.7472392","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7472392","url":null,"abstract":"A new algorithm called SUMIS-BO is proposed for soft-output MIMO detection. This method is a meaningful improvement of the \"Subspace Marginalization with Interference Suppression\" (SUMIS) algorithm. It exhibits good performance with reduced complexity and has been evaluated and compared in terms of performance and efficiency with the SUMIS algorithm using different system parameters. Results show that the performance of the SUMIS-BO is similar to the SUMIS algorithm, however its efficiency is improved. The new algorithm is far more efficient than SUMIS, especially with large systems.","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127443079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Wavelet-based decomposition of F0 as a secondary task for DNN-based speech synthesis with multi-task learning 基于小波分解的F0作为多任务学习的基于dnn的语音合成的辅助任务
M. Ribeiro, O. Watts, J. Yamagishi, R. Clark
We investigate two wavelet-based decomposition strategies of the f0 signal and their usefulness as a secondary task for speech synthesis using multi-task deep neural networks (MTL-DNN). The first decomposition strategy uses a static set of scales for all utterances in the training data. We propose a second strategy, where the scale of the mother wavelet is dynamically adjusted to the rate of each utterance. This approach is able to capture f0 variations related to the syllable, word, clitic-group, and phrase units. This method also constrains the wavelet components to be within the frequency range that previous experiments have shown to be more natural. These two strategies are evaluated as a secondary task in multi-task deep neural networks (MTL-DNNs). Results indicate that on an expressive dataset there is a strong preference for the systems using multi-task learning when compared to the baseline system.
我们研究了两种基于小波的f0信号分解策略,以及它们作为使用多任务深度神经网络(MTL-DNN)进行语音合成的辅助任务的有效性。第一种分解策略对训练数据中的所有话语使用一组静态尺度。我们提出了第二种策略,其中母小波的尺度根据每个话语的频率动态调整。这种方法能够捕获与音节、单词、关键字组和短语单位相关的60种变体。该方法还将小波分量限制在先前实验显示的更自然的频率范围内。在多任务深度神经网络(mtl - dnn)中,这两种策略作为次要任务进行评估。结果表明,与基线系统相比,在表达性数据集上,使用多任务学习的系统有强烈的偏好。
{"title":"Wavelet-based decomposition of F0 as a secondary task for DNN-based speech synthesis with multi-task learning","authors":"M. Ribeiro, O. Watts, J. Yamagishi, R. Clark","doi":"10.1109/ICASSP.2016.7472734","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7472734","url":null,"abstract":"We investigate two wavelet-based decomposition strategies of the f0 signal and their usefulness as a secondary task for speech synthesis using multi-task deep neural networks (MTL-DNN). The first decomposition strategy uses a static set of scales for all utterances in the training data. We propose a second strategy, where the scale of the mother wavelet is dynamically adjusted to the rate of each utterance. This approach is able to capture f0 variations related to the syllable, word, clitic-group, and phrase units. This method also constrains the wavelet components to be within the frequency range that previous experiments have shown to be more natural. These two strategies are evaluated as a secondary task in multi-task deep neural networks (MTL-DNNs). Results indicate that on an expressive dataset there is a strong preference for the systems using multi-task learning when compared to the baseline system.","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125095420","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Low-complexity recursive convolutional precoding for OFDM-based large-scale antenna systems 基于ofdm的大规模天线系统的低复杂度递归卷积预编码
Yinsheng Liu, Geoffrey Ye Li
Large-scale antenna (LSA) has gained a lot of attention recently since it can significantly improve the performance of wireless systems. Similar to multiple-input multiple-output (MIMO) orthogonal frequency division multiplexing (OFDM) or MIMO-OFDM, LSA can be also combined with OFDM to deal with frequency selectivity in wireless channels. However, such combination suffers from substantially increased complexity proportional to the number of antennas in LSA systems. In this paper, we propose a low-complexity recursive convolutional pre-coding to address the issues above. The traditional ZF precoding is implemented through the recursive convolutional precoding in the time domain so that only one IFFT is required for each user and the matrix inversion can be also avoided. Simulation results show that the proposed approach can achieve the same performance as that of ZF but with much lower complexity.
大型天线(large - large antenna, LSA)由于能够显著提高无线系统的性能,近年来受到了广泛的关注。与多输入多输出(MIMO)正交频分复用(OFDM)或MIMO-OFDM类似,LSA也可以与OFDM相结合来处理无线信道中的频率选择性。然而,这种组合的复杂性与LSA系统中天线的数量成正比。在本文中,我们提出了一种低复杂度递归卷积预编码来解决上述问题。传统的ZF预编码是通过时域递归卷积预编码来实现的,这样每个用户只需要一个IFFT,也可以避免矩阵反转。仿真结果表明,该方法可以达到与ZF算法相同的性能,但复杂度要低得多。
{"title":"Low-complexity recursive convolutional precoding for OFDM-based large-scale antenna systems","authors":"Yinsheng Liu, Geoffrey Ye Li","doi":"10.1109/ICASSP.2016.7472303","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7472303","url":null,"abstract":"Large-scale antenna (LSA) has gained a lot of attention recently since it can significantly improve the performance of wireless systems. Similar to multiple-input multiple-output (MIMO) orthogonal frequency division multiplexing (OFDM) or MIMO-OFDM, LSA can be also combined with OFDM to deal with frequency selectivity in wireless channels. However, such combination suffers from substantially increased complexity proportional to the number of antennas in LSA systems. In this paper, we propose a low-complexity recursive convolutional pre-coding to address the issues above. The traditional ZF precoding is implemented through the recursive convolutional precoding in the time domain so that only one IFFT is required for each user and the matrix inversion can be also avoided. Simulation results show that the proposed approach can achieve the same performance as that of ZF but with much lower complexity.","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126049737","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Long short term memory recurrent neural network based encoding method for emotion recognition in video 基于长短期记忆递归神经网络的视频情感识别编码方法
Linlin Chao, J. Tao, Minghao Yang, Ya Li, Zhengqi Wen
Human emotion is a temporally dynamic event which can be inferred from both audio and video feature sequences. In this paper we investigate the long short term memory recurrent neural network (LSTM-RNN) based encoding method for category emotion recognition in the video. LSTM-RNN is able to incorporate knowledge about how emotion evolves over long range successive frames and emotion clues from isolated frame. After encoding, each video clip can be represented by a vector for each input feature sequence. The vectors contain both frame level and sequence level emotion information. These vectors are then concatenated and fed into support vector machine (SVM) to get the final prediction result. Extensive evaluations on Emotion Challenge in the Wild (EmotiW2015) dataset show the efficiency of the proposed encoding method and competitive results are obtained. The final recognition accuracy achieves 46.38% for audio-video emotion recognition sub-challenge, where the challenge baseline is 39.33%.
人类情感是一种时间动态事件,可以从音频和视频特征序列中推断出来。本文研究了基于长短期记忆递归神经网络(LSTM-RNN)的视频分类情感识别编码方法。LSTM-RNN能够将情感如何在长距离连续帧中演变的知识和来自孤立帧的情感线索结合起来。编码后,每个视频片段可以用每个输入特征序列的向量表示。向量包含帧级和序列级情感信息。然后将这些向量连接并输入支持向量机(SVM)以获得最终的预测结果。对情感挑战(EmotiW2015)数据集的广泛评估表明了所提出的编码方法的有效性,并获得了具有竞争力的结果。音视频情感识别子挑战的最终识别准确率为46.38%,挑战基线为39.33%。
{"title":"Long short term memory recurrent neural network based encoding method for emotion recognition in video","authors":"Linlin Chao, J. Tao, Minghao Yang, Ya Li, Zhengqi Wen","doi":"10.1109/ICASSP.2016.7472178","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7472178","url":null,"abstract":"Human emotion is a temporally dynamic event which can be inferred from both audio and video feature sequences. In this paper we investigate the long short term memory recurrent neural network (LSTM-RNN) based encoding method for category emotion recognition in the video. LSTM-RNN is able to incorporate knowledge about how emotion evolves over long range successive frames and emotion clues from isolated frame. After encoding, each video clip can be represented by a vector for each input feature sequence. The vectors contain both frame level and sequence level emotion information. These vectors are then concatenated and fed into support vector machine (SVM) to get the final prediction result. Extensive evaluations on Emotion Challenge in the Wild (EmotiW2015) dataset show the efficiency of the proposed encoding method and competitive results are obtained. The final recognition accuracy achieves 46.38% for audio-video emotion recognition sub-challenge, where the challenge baseline is 39.33%.","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126102816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
One-bit ADCs in wideband massive MIMO systems with OFDM transmission OFDM传输的宽带大规模MIMO系统中的位adc
Christopher Mollén, Junil Choi, E. Larsson, R. Heath
We investigate the performance of wideband massive MIMO base stations that use one-bit ADCs for quantizing the uplink signal. Our main result is to show that the many taps of the frequency-selective channel make linear combiners asymptotically consistent and the quantization noise additive and Gaussian, which simplifies signal processing and enables the straightforward use of OFDM. We also find that single-carrier systems and OFDM systems are affected in the same way by one-bit quantizers in wideband systems because the distribution of the quantization noise becomes the same in both systems as the number of channel taps grows.
我们研究了使用1位adc量化上行信号的宽带大规模MIMO基站的性能。我们的主要结果是表明,频率选择信道的多个抽头使线性合成器渐近一致,量化噪声加性和高斯,这简化了信号处理并使OFDM的直接使用成为可能。我们还发现,单载波系统和OFDM系统在宽带系统中受到1位量化器的影响是相同的,因为随着信道抽头数量的增加,两种系统中量化噪声的分布变得相同。
{"title":"One-bit ADCs in wideband massive MIMO systems with OFDM transmission","authors":"Christopher Mollén, Junil Choi, E. Larsson, R. Heath","doi":"10.1109/ICASSP.2016.7472305","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7472305","url":null,"abstract":"We investigate the performance of wideband massive MIMO base stations that use one-bit ADCs for quantizing the uplink signal. Our main result is to show that the many taps of the frequency-selective channel make linear combiners asymptotically consistent and the quantization noise additive and Gaussian, which simplifies signal processing and enables the straightforward use of OFDM. We also find that single-carrier systems and OFDM systems are affected in the same way by one-bit quantizers in wideband systems because the distribution of the quantization noise becomes the same in both systems as the number of channel taps grows.","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123292206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
Single underwater image restoration by blue-green channels dehazing and red channel correction 采用蓝绿通道去雾和红通道校正的单幅水下图像恢复方法
Chongyi Li, Jichang Quo, Yanwei Pang, Shanji Chen, Jian Wang
Restoring underwater image from a single image is know to be ill-posed, and some assumptions made in previous methods are not suitable for many situations. In this paper, we propose a method based on blue-green channels dehazing and red channel correction for underwater image restoration. Firstly, blue-green channels are recovered via dehazing algorithm based on an extension and modification of Dark Channel Prior algorithm. Then, red channel is corrected following the Gray-World assumption theory. Finally, in order to resolve the problem which some recovered image regions may look too dim or too bright, an adaptive exposure map is built. Qualitative analysis demonstrates that our method significantly improves visibility and contrast, and reduces the effects of light absorption and scattering. For quantitative analysis, our results obtain best values in terms of entropy, local feature points and average gradient, which outperform three existing physical model available methods.
从单幅图像恢复水下图像是一种病态的方法,以前的方法所做的一些假设并不适用于许多情况。本文提出了一种基于蓝绿通道去雾和红通道校正的水下图像恢复方法。首先,在暗通道先验算法的基础上进行扩展和改进,通过去雾算法恢复蓝绿通道;然后,根据灰色世界假设理论对红色通道进行校正。最后,为了解决某些恢复图像区域看起来太暗或太亮的问题,构建了自适应曝光图。定性分析表明,我们的方法显著提高了能见度和对比度,减少了光吸收和散射的影响。在定量分析方面,我们的结果在熵、局部特征点和平均梯度方面获得了最佳值,优于现有的三种物理模型可用方法。
{"title":"Single underwater image restoration by blue-green channels dehazing and red channel correction","authors":"Chongyi Li, Jichang Quo, Yanwei Pang, Shanji Chen, Jian Wang","doi":"10.1109/ICASSP.2016.7471973","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7471973","url":null,"abstract":"Restoring underwater image from a single image is know to be ill-posed, and some assumptions made in previous methods are not suitable for many situations. In this paper, we propose a method based on blue-green channels dehazing and red channel correction for underwater image restoration. Firstly, blue-green channels are recovered via dehazing algorithm based on an extension and modification of Dark Channel Prior algorithm. Then, red channel is corrected following the Gray-World assumption theory. Finally, in order to resolve the problem which some recovered image regions may look too dim or too bright, an adaptive exposure map is built. Qualitative analysis demonstrates that our method significantly improves visibility and contrast, and reduces the effects of light absorption and scattering. For quantitative analysis, our results obtain best values in terms of entropy, local feature points and average gradient, which outperform three existing physical model available methods.","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"114 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123560293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 109
Capacity analysis of WCC-FBMC/OQAM systems WCC-FBMC/OQAM系统容量分析
Màrius Caus, A. Pérez-Neira, Adrian Kliks, Quentin Bodinier, F. Bader
This paper evaluates the capacity of weighted circularly convolved FBMC/OQAM systems. A rigorous mathematical model is derived to calculate the increase of capacity that can be obtained thanks to the lattice structure of the modulation and the exploitation of the intrinsic interference. The numerical results reveal that a signal to noise ratio gain of 2dB is obtained in one resource block of 12 subcarriers, translating into a capacity increase of 11% with respect to OFDM, which nuances some previous results on this topic in the literature.
本文评价了加权循环卷积FBMC/OQAM系统的容量。推导了一个严格的数学模型来计算由于调制的晶格结构和利用本征干涉所能获得的容量增加。数值结果表明,在12个子载波的一个资源块中获得了2dB的信噪比增益,相对于OFDM增加了11%的容量,这与文献中关于该主题的一些先前结果有细微差别。
{"title":"Capacity analysis of WCC-FBMC/OQAM systems","authors":"Màrius Caus, A. Pérez-Neira, Adrian Kliks, Quentin Bodinier, F. Bader","doi":"10.1109/ICASSP.2016.7472411","DOIUrl":"https://doi.org/10.1109/ICASSP.2016.7472411","url":null,"abstract":"This paper evaluates the capacity of weighted circularly convolved FBMC/OQAM systems. A rigorous mathematical model is derived to calculate the increase of capacity that can be obtained thanks to the lattice structure of the modulation and the exploitation of the intrinsic interference. The numerical results reveal that a signal to noise ratio gain of 2dB is obtained in one resource block of 12 subcarriers, translating into a capacity increase of 11% with respect to OFDM, which nuances some previous results on this topic in the literature.","PeriodicalId":165321,"journal":{"name":"2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125284656","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1